var/home/core/zuul-output/0000755000175000017500000000000015116236671014535 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116250001015460 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005375276415116247774017736 0ustar rootrootDec 10 09:33:56 crc systemd[1]: Starting Kubernetes Kubelet... Dec 10 09:33:56 crc restorecon[4683]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:56 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:57 crc restorecon[4683]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 09:33:57 crc restorecon[4683]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 10 09:33:57 crc kubenswrapper[4715]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 09:33:57 crc kubenswrapper[4715]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 10 09:33:57 crc kubenswrapper[4715]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 09:33:57 crc kubenswrapper[4715]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 09:33:57 crc kubenswrapper[4715]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 10 09:33:57 crc kubenswrapper[4715]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.401693 4715 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406239 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406265 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406273 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406285 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406293 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406301 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406317 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406324 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406330 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406359 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406369 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406379 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406388 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406396 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406403 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406410 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406418 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406425 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406473 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406528 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406540 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406550 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406560 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406569 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406578 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406589 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406598 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406608 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406617 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406625 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406644 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406655 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406665 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406673 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406683 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406693 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406702 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406712 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406733 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406751 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406761 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406771 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406788 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406797 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406806 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406816 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406826 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406835 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406844 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406852 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406863 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406872 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406881 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406889 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406907 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406948 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406957 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406966 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406975 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406984 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.406993 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.407001 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.407010 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.407019 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.407092 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.407356 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.407371 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.407381 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.407395 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.407407 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.407416 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407593 4715 flags.go:64] FLAG: --address="0.0.0.0" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407621 4715 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407644 4715 flags.go:64] FLAG: --anonymous-auth="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407661 4715 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407678 4715 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407692 4715 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407709 4715 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407725 4715 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407740 4715 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407753 4715 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407766 4715 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407780 4715 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407792 4715 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407805 4715 flags.go:64] FLAG: --cgroup-root="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407817 4715 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407831 4715 flags.go:64] FLAG: --client-ca-file="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407844 4715 flags.go:64] FLAG: --cloud-config="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407856 4715 flags.go:64] FLAG: --cloud-provider="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407869 4715 flags.go:64] FLAG: --cluster-dns="[]" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407887 4715 flags.go:64] FLAG: --cluster-domain="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407899 4715 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407954 4715 flags.go:64] FLAG: --config-dir="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407968 4715 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.407982 4715 flags.go:64] FLAG: --container-log-max-files="5" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408018 4715 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408031 4715 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408044 4715 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408057 4715 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408070 4715 flags.go:64] FLAG: --contention-profiling="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408087 4715 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408100 4715 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408113 4715 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408125 4715 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408142 4715 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408154 4715 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408164 4715 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408173 4715 flags.go:64] FLAG: --enable-load-reader="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408183 4715 flags.go:64] FLAG: --enable-server="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408193 4715 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408208 4715 flags.go:64] FLAG: --event-burst="100" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408219 4715 flags.go:64] FLAG: --event-qps="50" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408229 4715 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408238 4715 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408248 4715 flags.go:64] FLAG: --eviction-hard="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408260 4715 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408270 4715 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408280 4715 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408291 4715 flags.go:64] FLAG: --eviction-soft="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408301 4715 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408311 4715 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408321 4715 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408331 4715 flags.go:64] FLAG: --experimental-mounter-path="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408340 4715 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408350 4715 flags.go:64] FLAG: --fail-swap-on="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408359 4715 flags.go:64] FLAG: --feature-gates="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408373 4715 flags.go:64] FLAG: --file-check-frequency="20s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408383 4715 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408393 4715 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408403 4715 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408413 4715 flags.go:64] FLAG: --healthz-port="10248" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408424 4715 flags.go:64] FLAG: --help="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408434 4715 flags.go:64] FLAG: --hostname-override="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408443 4715 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408453 4715 flags.go:64] FLAG: --http-check-frequency="20s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408463 4715 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408473 4715 flags.go:64] FLAG: --image-credential-provider-config="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408484 4715 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408497 4715 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408508 4715 flags.go:64] FLAG: --image-service-endpoint="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408520 4715 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408532 4715 flags.go:64] FLAG: --kube-api-burst="100" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408546 4715 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408562 4715 flags.go:64] FLAG: --kube-api-qps="50" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408574 4715 flags.go:64] FLAG: --kube-reserved="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408588 4715 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408600 4715 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408613 4715 flags.go:64] FLAG: --kubelet-cgroups="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408626 4715 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408639 4715 flags.go:64] FLAG: --lock-file="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408653 4715 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408666 4715 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408680 4715 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408699 4715 flags.go:64] FLAG: --log-json-split-stream="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408709 4715 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408719 4715 flags.go:64] FLAG: --log-text-split-stream="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408731 4715 flags.go:64] FLAG: --logging-format="text" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408744 4715 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408759 4715 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408772 4715 flags.go:64] FLAG: --manifest-url="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408785 4715 flags.go:64] FLAG: --manifest-url-header="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408803 4715 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408816 4715 flags.go:64] FLAG: --max-open-files="1000000" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408833 4715 flags.go:64] FLAG: --max-pods="110" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408846 4715 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408860 4715 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408872 4715 flags.go:64] FLAG: --memory-manager-policy="None" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408885 4715 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408899 4715 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408947 4715 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408961 4715 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408989 4715 flags.go:64] FLAG: --node-status-max-images="50" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.408999 4715 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409012 4715 flags.go:64] FLAG: --oom-score-adj="-999" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409025 4715 flags.go:64] FLAG: --pod-cidr="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409038 4715 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409060 4715 flags.go:64] FLAG: --pod-manifest-path="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409071 4715 flags.go:64] FLAG: --pod-max-pids="-1" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409081 4715 flags.go:64] FLAG: --pods-per-core="0" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409091 4715 flags.go:64] FLAG: --port="10250" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409102 4715 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409115 4715 flags.go:64] FLAG: --provider-id="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409129 4715 flags.go:64] FLAG: --qos-reserved="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409145 4715 flags.go:64] FLAG: --read-only-port="10255" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409185 4715 flags.go:64] FLAG: --register-node="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409198 4715 flags.go:64] FLAG: --register-schedulable="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409211 4715 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409235 4715 flags.go:64] FLAG: --registry-burst="10" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409249 4715 flags.go:64] FLAG: --registry-qps="5" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409262 4715 flags.go:64] FLAG: --reserved-cpus="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409278 4715 flags.go:64] FLAG: --reserved-memory="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409295 4715 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409308 4715 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409323 4715 flags.go:64] FLAG: --rotate-certificates="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409335 4715 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409348 4715 flags.go:64] FLAG: --runonce="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409360 4715 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409373 4715 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409387 4715 flags.go:64] FLAG: --seccomp-default="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409399 4715 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409411 4715 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409425 4715 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409437 4715 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409450 4715 flags.go:64] FLAG: --storage-driver-password="root" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409463 4715 flags.go:64] FLAG: --storage-driver-secure="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409474 4715 flags.go:64] FLAG: --storage-driver-table="stats" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409484 4715 flags.go:64] FLAG: --storage-driver-user="root" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409494 4715 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409504 4715 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409514 4715 flags.go:64] FLAG: --system-cgroups="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409526 4715 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409564 4715 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409580 4715 flags.go:64] FLAG: --tls-cert-file="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409593 4715 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409615 4715 flags.go:64] FLAG: --tls-min-version="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409628 4715 flags.go:64] FLAG: --tls-private-key-file="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409640 4715 flags.go:64] FLAG: --topology-manager-policy="none" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409652 4715 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409665 4715 flags.go:64] FLAG: --topology-manager-scope="container" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409678 4715 flags.go:64] FLAG: --v="2" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409696 4715 flags.go:64] FLAG: --version="false" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409710 4715 flags.go:64] FLAG: --vmodule="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409721 4715 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.409732 4715 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410012 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410024 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410035 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410044 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410052 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410060 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410072 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410082 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410091 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410103 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410114 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410124 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410133 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410142 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410151 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410161 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410170 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410179 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410187 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410195 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410204 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410212 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410221 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410230 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410238 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410246 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410254 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410263 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410272 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410280 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410288 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410297 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410305 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410314 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410322 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410332 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410340 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410349 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410358 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410367 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410375 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410383 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410392 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410400 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410408 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410416 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410424 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410432 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410441 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410450 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410459 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410467 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410476 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410484 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410493 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410502 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410511 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410519 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410528 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410536 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410544 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410553 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410564 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410575 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410586 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410596 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410606 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410615 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410624 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410634 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.410643 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.410944 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.422896 4715 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.422979 4715 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423118 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423134 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423146 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423155 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423166 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423176 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423185 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423194 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423202 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423211 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423220 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423228 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423237 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423245 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423254 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423262 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423271 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423280 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423289 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423298 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423307 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423318 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423327 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423336 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423345 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423353 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423366 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423379 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423391 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423401 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423410 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423419 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423427 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423437 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423447 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423457 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423468 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423479 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423490 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423501 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423510 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423519 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423528 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423538 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423546 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423555 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423564 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423573 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423582 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423591 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423599 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423607 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423616 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423625 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423634 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423643 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423651 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423660 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423668 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423677 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423686 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423695 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423703 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423712 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423720 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423729 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423738 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423747 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423757 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423765 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.423778 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.423793 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424081 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424099 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424110 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424120 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424129 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424137 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424146 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424155 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424165 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424173 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424181 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424189 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424198 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424207 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424218 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424233 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424243 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424254 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424264 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424273 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424284 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424293 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424302 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424311 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424320 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424329 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424337 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424346 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424354 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424363 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424372 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424380 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424388 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424400 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424413 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424424 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424433 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424442 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424451 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424460 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424468 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424476 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424484 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424493 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424503 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424512 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424521 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424529 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424538 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424546 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424554 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424562 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424571 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424580 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424588 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424597 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424606 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424614 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424623 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424631 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424639 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424648 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424657 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424666 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424674 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424683 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424692 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424701 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424712 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424722 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.424732 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.424746 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.425099 4715 server.go:940] "Client rotation is on, will bootstrap in background" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.430138 4715 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.430346 4715 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.437805 4715 server.go:997] "Starting client certificate rotation" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.437888 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.438141 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-04 21:54:43.941486893 +0000 UTC Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.438254 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.453575 4715 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.455072 4715 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.456329 4715 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.467488 4715 log.go:25] "Validated CRI v1 runtime API" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.480976 4715 log.go:25] "Validated CRI v1 image API" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.482476 4715 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.485719 4715 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-10-09-29-41-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.485951 4715 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.510460 4715 manager.go:217] Machine: {Timestamp:2025-12-10 09:33:57.507989425 +0000 UTC m=+0.251535766 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:13f37a17-0606-45c2-b39f-55bf2b8fcb59 BootID:29a4a427-b049-4b60-be94-f8a6c917cb86 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:91:c6:12 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:91:c6:12 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:62:f8:2d Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:b9:35:6a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:79:11:88 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:8b:7b:14 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4e:8b:c9:f6:36:35 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:92:cb:ea:eb:82:a9 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.511082 4715 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.511457 4715 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.512346 4715 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.512856 4715 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.513094 4715 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.513732 4715 topology_manager.go:138] "Creating topology manager with none policy" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.513770 4715 container_manager_linux.go:303] "Creating device plugin manager" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.514194 4715 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.514276 4715 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.514866 4715 state_mem.go:36] "Initialized new in-memory state store" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.515121 4715 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.530342 4715 kubelet.go:418] "Attempting to sync node with API server" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.530370 4715 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.530398 4715 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.530411 4715 kubelet.go:324] "Adding apiserver pod source" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.530424 4715 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.532732 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.532793 4715 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.532740 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.532947 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.532983 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.533302 4715 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.533978 4715 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534668 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534696 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534705 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534714 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534728 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534738 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534747 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534761 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534771 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534781 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534793 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.534802 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.535005 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.535470 4715 server.go:1280] "Started kubelet" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.535995 4715 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.535959 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.537093 4715 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 10 09:33:57 crc systemd[1]: Started Kubernetes Kubelet. Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.539142 4715 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.539274 4715 server.go:460] "Adding debug handlers to kubelet server" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.540416 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.541256 4715 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.541417 4715 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.541440 4715 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.542134 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 16:55:57.000648013 +0000 UTC Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.542300 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 439h21m59.458353049s for next certificate rotation Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.543500 4715 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.543602 4715 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.541269 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.201:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187fd0df8e410f88 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 09:33:57.53543668 +0000 UTC m=+0.278982941,LastTimestamp:2025-12-10 09:33:57.53543668 +0000 UTC m=+0.278982941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.544214 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.544300 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="200ms" Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.544357 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.552054 4715 factory.go:55] Registering systemd factory Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.552142 4715 factory.go:221] Registration of the systemd container factory successfully Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.553967 4715 factory.go:153] Registering CRI-O factory Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.554013 4715 factory.go:221] Registration of the crio container factory successfully Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.554125 4715 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.554170 4715 factory.go:103] Registering Raw factory Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.554199 4715 manager.go:1196] Started watching for new ooms in manager Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.555498 4715 manager.go:319] Starting recovery of all containers Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.559806 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.559877 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.559891 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.559904 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.559935 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.559947 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.559959 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.559970 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.559984 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.559996 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560007 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560018 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560029 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560044 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560058 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560068 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560082 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560093 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560107 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560120 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560157 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560170 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560181 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560193 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560210 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560224 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560241 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560299 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560316 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560349 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560390 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560406 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560423 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560438 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560451 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560465 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560478 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560493 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560505 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560520 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560533 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560547 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560560 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560573 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560587 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560599 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560612 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560626 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560639 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560652 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560664 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560676 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560693 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560707 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560720 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560734 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560746 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560757 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560769 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560781 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560792 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560808 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560823 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560837 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560849 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560862 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560873 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560884 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560896 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.560995 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561009 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561021 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561032 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561043 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561057 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561072 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561084 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561096 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561146 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561157 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561169 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561188 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561199 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561212 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561224 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561235 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561247 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561260 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561272 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561283 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561297 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561310 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561322 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561334 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561346 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561357 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561368 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561380 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561392 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561406 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561418 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561430 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561443 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561457 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561475 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561489 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561507 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561526 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561539 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561551 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561564 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561576 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561590 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561604 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561616 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561628 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561641 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561654 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561666 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561680 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561692 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561705 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561718 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561731 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561745 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561759 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561771 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561783 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561796 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561808 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561821 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561832 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561845 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561856 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561869 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561881 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561894 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561905 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561936 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561951 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561964 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561976 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.561988 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562000 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562013 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562025 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562038 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562050 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562061 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562073 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562085 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562096 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562107 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562120 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562131 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562143 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562155 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562168 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562179 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562191 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562203 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562214 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562226 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562237 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562249 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562261 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562273 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562285 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562297 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562309 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562321 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562337 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562348 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562360 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562411 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562424 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562439 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562451 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562463 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562474 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562486 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562498 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562511 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.562524 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565491 4715 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565541 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565568 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565586 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565605 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565630 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565652 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565670 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565686 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565704 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565719 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565735 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565754 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565771 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565790 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565807 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565824 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565843 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565865 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565883 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565899 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565937 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565958 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565976 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.565991 4715 reconstruct.go:97] "Volume reconstruction finished" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.566002 4715 reconciler.go:26] "Reconciler: start to sync state" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.594286 4715 manager.go:324] Recovery completed Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.601703 4715 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.603194 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.603590 4715 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.603624 4715 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.603646 4715 kubelet.go:2335] "Starting kubelet main sync loop" Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.603682 4715 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 10 09:33:57 crc kubenswrapper[4715]: W1210 09:33:57.604600 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.604670 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.605814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.605849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.605859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.606592 4715 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.606609 4715 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.606628 4715 state_mem.go:36] "Initialized new in-memory state store" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.631905 4715 policy_none.go:49] "None policy: Start" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.632601 4715 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.632626 4715 state_mem.go:35] "Initializing new in-memory state store" Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.644642 4715 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.687954 4715 manager.go:334] "Starting Device Plugin manager" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.688002 4715 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.688014 4715 server.go:79] "Starting device plugin registration server" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.688380 4715 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.688398 4715 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.688692 4715 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.688761 4715 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.688768 4715 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.695383 4715 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.704605 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.704719 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.705603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.705634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.705644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.705743 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.705953 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.705984 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.706352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.706405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.706421 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.706479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.706499 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.706508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.706618 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.706755 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.706805 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.707403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.707429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.707439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.707551 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.707562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.707575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.707585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.707874 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.707926 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708041 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708103 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708285 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708306 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708898 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.708937 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.709355 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.709374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.709382 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.709451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.709466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.709473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.745391 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="400ms" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.767110 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.767160 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.767230 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.767333 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.767711 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.767962 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.768001 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.768035 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.768068 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.768100 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.768132 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.768160 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.768215 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.768326 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.768425 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.788813 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.790223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.790293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.790314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.790351 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.791370 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.201:6443: connect: connection refused" node="crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870054 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870107 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870124 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870143 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870161 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870179 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870197 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870214 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870231 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870252 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870245 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870282 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870298 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870330 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870274 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870249 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870345 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870301 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870368 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870320 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870448 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870464 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870508 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870500 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870554 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870525 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870571 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870583 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870609 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.870731 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.991564 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.992869 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.992899 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.992909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:57 crc kubenswrapper[4715]: I1210 09:33:57.992942 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 09:33:57 crc kubenswrapper[4715]: E1210 09:33:57.993351 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.201:6443: connect: connection refused" node="crc" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.037095 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.061330 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 10 09:33:58 crc kubenswrapper[4715]: W1210 09:33:58.076044 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-8148f1ee282473fbb552d21ecc1217fea2e0ec6ffc1005352b7413a3d1d1484a WatchSource:0}: Error finding container 8148f1ee282473fbb552d21ecc1217fea2e0ec6ffc1005352b7413a3d1d1484a: Status 404 returned error can't find the container with id 8148f1ee282473fbb552d21ecc1217fea2e0ec6ffc1005352b7413a3d1d1484a Dec 10 09:33:58 crc kubenswrapper[4715]: W1210 09:33:58.084629 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-5f86d86f917591d8ca198de886158c2e7145d9a14ca814e48d399d3686ebee3c WatchSource:0}: Error finding container 5f86d86f917591d8ca198de886158c2e7145d9a14ca814e48d399d3686ebee3c: Status 404 returned error can't find the container with id 5f86d86f917591d8ca198de886158c2e7145d9a14ca814e48d399d3686ebee3c Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.084894 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.091653 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.094624 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 09:33:58 crc kubenswrapper[4715]: W1210 09:33:58.115808 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-78078e7bdeb8154a8de1927ad5e9db2132ffbcaae633637209414d6aa196d61f WatchSource:0}: Error finding container 78078e7bdeb8154a8de1927ad5e9db2132ffbcaae633637209414d6aa196d61f: Status 404 returned error can't find the container with id 78078e7bdeb8154a8de1927ad5e9db2132ffbcaae633637209414d6aa196d61f Dec 10 09:33:58 crc kubenswrapper[4715]: W1210 09:33:58.118069 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-3a9ee8245b205fc4b214061b2658ef47e48c71f246aec5d3b916f8215401a412 WatchSource:0}: Error finding container 3a9ee8245b205fc4b214061b2658ef47e48c71f246aec5d3b916f8215401a412: Status 404 returned error can't find the container with id 3a9ee8245b205fc4b214061b2658ef47e48c71f246aec5d3b916f8215401a412 Dec 10 09:33:58 crc kubenswrapper[4715]: W1210 09:33:58.132017 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-7a5eceb40991cda9d2e2e8e60f8efe2f327967a3a6975c503118bddf8e454b75 WatchSource:0}: Error finding container 7a5eceb40991cda9d2e2e8e60f8efe2f327967a3a6975c503118bddf8e454b75: Status 404 returned error can't find the container with id 7a5eceb40991cda9d2e2e8e60f8efe2f327967a3a6975c503118bddf8e454b75 Dec 10 09:33:58 crc kubenswrapper[4715]: E1210 09:33:58.146214 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="800ms" Dec 10 09:33:58 crc kubenswrapper[4715]: W1210 09:33:58.373272 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 09:33:58 crc kubenswrapper[4715]: E1210 09:33:58.373359 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.394051 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.395422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.395459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.395472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.395497 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 09:33:58 crc kubenswrapper[4715]: E1210 09:33:58.395999 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.201:6443: connect: connection refused" node="crc" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.537076 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 09:33:58 crc kubenswrapper[4715]: W1210 09:33:58.552774 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 09:33:58 crc kubenswrapper[4715]: E1210 09:33:58.552860 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 09:33:58 crc kubenswrapper[4715]: W1210 09:33:58.581868 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 09:33:58 crc kubenswrapper[4715]: E1210 09:33:58.581979 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.610251 4715 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb" exitCode=0 Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.610324 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb"} Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.610466 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7a5eceb40991cda9d2e2e8e60f8efe2f327967a3a6975c503118bddf8e454b75"} Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.610678 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.611812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.611843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.611855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.612427 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667"} Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.612494 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3a9ee8245b205fc4b214061b2658ef47e48c71f246aec5d3b916f8215401a412"} Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.614105 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9" exitCode=0 Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.614145 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9"} Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.614161 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"78078e7bdeb8154a8de1927ad5e9db2132ffbcaae633637209414d6aa196d61f"} Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.614241 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.614820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.614865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.614878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.666134 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.666999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.667026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.667036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.668560 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03" exitCode=0 Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.668633 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03"} Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.668671 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5f86d86f917591d8ca198de886158c2e7145d9a14ca814e48d399d3686ebee3c"} Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.668768 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.669620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.669663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.669676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.670648 4715 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a" exitCode=0 Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.670685 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a"} Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.670717 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8148f1ee282473fbb552d21ecc1217fea2e0ec6ffc1005352b7413a3d1d1484a"} Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.670814 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.671564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.671595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:58 crc kubenswrapper[4715]: I1210 09:33:58.671605 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:58 crc kubenswrapper[4715]: W1210 09:33:58.783831 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.201:6443: connect: connection refused Dec 10 09:33:58 crc kubenswrapper[4715]: E1210 09:33:58.783904 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.201:6443: connect: connection refused" logger="UnhandledError" Dec 10 09:33:58 crc kubenswrapper[4715]: E1210 09:33:58.947196 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="1.6s" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.196272 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.203351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.203402 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.203411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.203435 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 09:33:59 crc kubenswrapper[4715]: E1210 09:33:59.204393 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.201:6443: connect: connection refused" node="crc" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.653779 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.674901 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4645809ac50eb70b0ad1ba89039994d876bc7114e3d40ba800e49d171f8f3fa7"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.675087 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.676318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.676365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.676375 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.677902 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.677948 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.677958 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.678030 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.678741 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.678781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.678796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.680756 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.680776 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.680786 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.680800 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.681477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.681500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.681510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.686000 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.686030 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.686041 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.686050 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.686059 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.686130 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.686817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.686845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.686853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.689144 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f" exitCode=0 Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.689198 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f"} Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.689354 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.690359 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.690384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:33:59 crc kubenswrapper[4715]: I1210 09:33:59.690393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.693780 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4" exitCode=0 Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.693873 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4"} Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.694591 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.694778 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.695865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.695898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.695927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.696014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.696046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.696059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.805477 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.806681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.806727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.806738 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.806767 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.826635 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.826813 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.827973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.828021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:00 crc kubenswrapper[4715]: I1210 09:34:00.828030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.104984 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.105178 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.105219 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.106335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.106404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.106428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.548790 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.701347 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565"} Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.701398 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db"} Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.701414 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd"} Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.701426 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30"} Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.701433 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.703426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.703479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:01 crc kubenswrapper[4715]: I1210 09:34:01.703496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:02 crc kubenswrapper[4715]: I1210 09:34:02.709054 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c"} Dec 10 09:34:02 crc kubenswrapper[4715]: I1210 09:34:02.709176 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:02 crc kubenswrapper[4715]: I1210 09:34:02.710040 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:02 crc kubenswrapper[4715]: I1210 09:34:02.710067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:02 crc kubenswrapper[4715]: I1210 09:34:02.710074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:02 crc kubenswrapper[4715]: I1210 09:34:02.951631 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:34:02 crc kubenswrapper[4715]: I1210 09:34:02.951839 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:02 crc kubenswrapper[4715]: I1210 09:34:02.953553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:02 crc kubenswrapper[4715]: I1210 09:34:02.953603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:02 crc kubenswrapper[4715]: I1210 09:34:02.953620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:03 crc kubenswrapper[4715]: I1210 09:34:03.342254 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:34:03 crc kubenswrapper[4715]: I1210 09:34:03.342662 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:34:03 crc kubenswrapper[4715]: I1210 09:34:03.342752 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:03 crc kubenswrapper[4715]: I1210 09:34:03.343852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:03 crc kubenswrapper[4715]: I1210 09:34:03.343931 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:03 crc kubenswrapper[4715]: I1210 09:34:03.343951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:03 crc kubenswrapper[4715]: I1210 09:34:03.711958 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:03 crc kubenswrapper[4715]: I1210 09:34:03.713329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:03 crc kubenswrapper[4715]: I1210 09:34:03.713387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:03 crc kubenswrapper[4715]: I1210 09:34:03.713424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:04 crc kubenswrapper[4715]: I1210 09:34:04.549690 4715 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 09:34:04 crc kubenswrapper[4715]: I1210 09:34:04.549803 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.615432 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.615706 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.617055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.617092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.617103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.780666 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.780882 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.782551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.782596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.782604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.787726 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.851731 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.852007 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.853677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.853732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:06 crc kubenswrapper[4715]: I1210 09:34:06.853741 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:07 crc kubenswrapper[4715]: E1210 09:34:07.695606 4715 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 09:34:07 crc kubenswrapper[4715]: I1210 09:34:07.723279 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:07 crc kubenswrapper[4715]: I1210 09:34:07.724779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:07 crc kubenswrapper[4715]: I1210 09:34:07.724836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:07 crc kubenswrapper[4715]: I1210 09:34:07.724846 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:07 crc kubenswrapper[4715]: I1210 09:34:07.882654 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:34:07 crc kubenswrapper[4715]: I1210 09:34:07.882985 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:07 crc kubenswrapper[4715]: I1210 09:34:07.884376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:07 crc kubenswrapper[4715]: I1210 09:34:07.884415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:07 crc kubenswrapper[4715]: I1210 09:34:07.884426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:09 crc kubenswrapper[4715]: I1210 09:34:09.054354 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 10 09:34:09 crc kubenswrapper[4715]: I1210 09:34:09.055839 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:09 crc kubenswrapper[4715]: I1210 09:34:09.057465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:09 crc kubenswrapper[4715]: I1210 09:34:09.057502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:09 crc kubenswrapper[4715]: I1210 09:34:09.057518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:09 crc kubenswrapper[4715]: I1210 09:34:09.537886 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 10 09:34:09 crc kubenswrapper[4715]: E1210 09:34:09.655484 4715 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 10 09:34:10 crc kubenswrapper[4715]: I1210 09:34:10.113393 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 10 09:34:10 crc kubenswrapper[4715]: I1210 09:34:10.113499 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 10 09:34:10 crc kubenswrapper[4715]: I1210 09:34:10.121999 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 10 09:34:10 crc kubenswrapper[4715]: I1210 09:34:10.122073 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.346815 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.347009 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.348038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.348077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.348088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.351323 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.739197 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.740732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.740788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.740809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.780824 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 09:34:13 crc kubenswrapper[4715]: I1210 09:34:13.798534 4715 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 10 09:34:14 crc kubenswrapper[4715]: I1210 09:34:14.537543 4715 csr.go:261] certificate signing request csr-bg6tf is approved, waiting to be issued Dec 10 09:34:14 crc kubenswrapper[4715]: I1210 09:34:14.550001 4715 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 09:34:14 crc kubenswrapper[4715]: I1210 09:34:14.550075 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 09:34:14 crc kubenswrapper[4715]: I1210 09:34:14.551925 4715 csr.go:257] certificate signing request csr-bg6tf is issued Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.112275 4715 trace.go:236] Trace[524007169]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 09:34:00.245) (total time: 14866ms): Dec 10 09:34:15 crc kubenswrapper[4715]: Trace[524007169]: ---"Objects listed" error: 14866ms (09:34:15.112) Dec 10 09:34:15 crc kubenswrapper[4715]: Trace[524007169]: [14.866940833s] [14.866940833s] END Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.112309 4715 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.112322 4715 trace.go:236] Trace[720321346]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 09:34:00.935) (total time: 14176ms): Dec 10 09:34:15 crc kubenswrapper[4715]: Trace[720321346]: ---"Objects listed" error: 14176ms (09:34:15.112) Dec 10 09:34:15 crc kubenswrapper[4715]: Trace[720321346]: [14.17692742s] [14.17692742s] END Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.112359 4715 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.112579 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.115370 4715 trace.go:236] Trace[1737326664]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 09:34:01.564) (total time: 13550ms): Dec 10 09:34:15 crc kubenswrapper[4715]: Trace[1737326664]: ---"Objects listed" error: 13550ms (09:34:15.115) Dec 10 09:34:15 crc kubenswrapper[4715]: Trace[1737326664]: [13.550439343s] [13.550439343s] END Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.115587 4715 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.115597 4715 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.116347 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.116626 4715 trace.go:236] Trace[741429860]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 09:34:01.594) (total time: 13521ms): Dec 10 09:34:15 crc kubenswrapper[4715]: Trace[741429860]: ---"Objects listed" error: 13521ms (09:34:15.116) Dec 10 09:34:15 crc kubenswrapper[4715]: Trace[741429860]: [13.521866918s] [13.521866918s] END Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.116665 4715 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.392702 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:34446->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.392814 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:34446->192.168.126.11:17697: read: connection reset by peer" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.393510 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.393595 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.542363 4715 apiserver.go:52] "Watching apiserver" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.545357 4715 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.545713 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.546224 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.546291 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.546395 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.546419 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.546669 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.546741 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.546879 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.547053 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.547116 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.548356 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.548582 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.549152 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.549239 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.549371 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.550783 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.552069 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.555124 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-10 09:29:14 +0000 UTC, rotation deadline is 2026-10-20 11:58:58.287961592 +0000 UTC Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.555158 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7538h24m42.732807478s for next certificate rotation Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.555214 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.555338 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.575596 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.594498 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.603960 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.615811 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.627576 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.637732 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.644653 4715 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.649950 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.660875 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.672184 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719673 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719750 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719769 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719791 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719816 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719834 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719855 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719874 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719906 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719954 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719969 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.719990 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720006 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720021 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720039 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720055 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720071 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720091 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720106 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720123 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720139 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720154 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720217 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720235 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720246 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720286 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720271 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720336 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720356 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720365 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720372 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720419 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720443 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720464 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720482 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720488 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720499 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720516 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720536 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720556 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720575 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720594 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720618 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720640 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720624 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720657 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720677 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720741 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720771 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720777 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720796 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720802 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720817 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720841 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720862 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720884 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720905 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720922 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720947 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720971 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.720979 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721017 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721025 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721023 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721049 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721061 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721086 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721119 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721134 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721145 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721169 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721176 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721197 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721219 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721246 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721270 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721295 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721317 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721320 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721323 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721341 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721366 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721390 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721415 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721444 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721446 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721470 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721496 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721521 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721547 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721569 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721591 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721614 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721638 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721661 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721869 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721898 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721985 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722021 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722047 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722070 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722094 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722150 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722208 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722307 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722333 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722357 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.723818 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.723844 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.724119 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.724425 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.724766 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721493 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721532 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721627 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721648 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721689 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721734 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721937 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.721997 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722118 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722215 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722417 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.722976 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.723038 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.723607 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.723628 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.723627 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.723730 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.723990 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.723977 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.724576 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.724741 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.724757 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.724818 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.725105 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.725143 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.725525 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.725662 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727601 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727627 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727635 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727775 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727813 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727842 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727841 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727867 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727893 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727936 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727939 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727960 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.727989 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728029 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728133 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728149 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728163 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728190 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728311 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728347 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728373 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728400 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728419 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728462 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728490 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728539 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728592 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728618 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728645 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728663 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728672 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728686 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728764 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728800 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728838 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728825 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728866 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728929 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728957 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.728981 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729002 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729023 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729042 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729049 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729063 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729073 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729289 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729397 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729765 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729818 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.729958 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730052 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730094 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730122 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730146 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730150 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730171 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730201 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730311 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730208 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730211 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730361 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730452 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730500 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730531 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730560 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730598 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730634 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730659 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730684 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730711 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730740 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730764 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730789 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730812 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730841 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730869 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730900 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730948 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730950 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.730981 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731001 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731015 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731075 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731109 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731135 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731164 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731317 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731346 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731357 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731385 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731414 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731439 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731465 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731491 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731516 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731551 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731571 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731598 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731650 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731672 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731720 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731741 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731762 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731779 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731799 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731816 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731835 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731851 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731867 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731884 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731899 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731982 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732001 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732027 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732052 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732079 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732095 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732116 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732140 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732159 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732178 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732199 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732225 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733095 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733122 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733143 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733183 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733205 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733247 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733278 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733342 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733369 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733414 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733434 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733450 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733468 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733572 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733587 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733597 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733609 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733619 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733628 4715 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733637 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733647 4715 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733656 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733665 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733675 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733685 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733694 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733703 4715 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733713 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733721 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733730 4715 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733741 4715 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733751 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733762 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733772 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733782 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733791 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733802 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733812 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733821 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733832 4715 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733842 4715 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733851 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733861 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733875 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733883 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733892 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733904 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733925 4715 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733935 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733945 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733955 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733964 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733974 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733985 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733995 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734004 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734013 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734022 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734031 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734040 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734050 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734096 4715 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734107 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734118 4715 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734133 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734149 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734181 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734192 4715 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734211 4715 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734221 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734231 4715 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734243 4715 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734258 4715 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734270 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734283 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734292 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734302 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734311 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734320 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734335 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734347 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.740225 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.740552 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.740851 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731411 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731471 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731523 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.731722 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732251 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732292 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732308 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732379 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732505 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732573 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732658 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.732688 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733058 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733470 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733537 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733778 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.748043 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.748066 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.748080 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733863 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733948 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.733975 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734104 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734186 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734197 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734505 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734636 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.734873 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.736064 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:34:16.236028494 +0000 UTC m=+18.979574935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.748269 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:16.248247809 +0000 UTC m=+18.991794050 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.736131 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.736358 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.736572 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.736687 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.748338 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:16.248331291 +0000 UTC m=+18.991877532 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.736776 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.748368 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:16.248362942 +0000 UTC m=+18.991909193 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.737023 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.737034 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.737187 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.737210 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.737121 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.737427 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.737466 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.737502 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.737711 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.737845 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.738084 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.738124 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.738160 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.738300 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.738473 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.738594 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.738643 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.738731 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.738754 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.739022 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.739353 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.740003 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.740691 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.740886 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.741060 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.741270 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.741379 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.741782 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.741847 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.741939 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.741987 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.742439 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.742715 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.742718 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.742885 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.743043 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.743270 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.741125 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.744068 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.743900 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.745132 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.747401 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.748969 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.749247 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.749272 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.750030 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.750086 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.750356 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.750560 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.750935 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.751225 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.751316 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.751702 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.751744 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.752244 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.751969 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.752554 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.752845 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.752730 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.753025 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.753308 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.753370 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.753832 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.753844 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.753864 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.753876 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.754471 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.754525 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.754010 4715 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 10 09:34:15 crc kubenswrapper[4715]: E1210 09:34:15.755019 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:16.253899723 +0000 UTC m=+18.997445974 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.754432 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.756045 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.755978 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.756157 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.756136 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.756237 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.757580 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.759292 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.759884 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.760036 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.759897 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.760341 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.760490 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.760623 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.760669 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.760675 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.760747 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.760769 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.760943 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.760976 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.761085 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.761508 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.763109 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.763184 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.763290 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.763550 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.763665 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.763704 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47" exitCode=255 Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.763738 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47"} Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.764720 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.778099 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.778302 4715 scope.go:117] "RemoveContainer" containerID="6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.778297 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.778764 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.780823 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.784886 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.784951 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.785255 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.785593 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.785802 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.786179 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.791906 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.796869 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.804622 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.808536 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.815754 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.816092 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.829710 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835124 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835166 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835250 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835264 4715 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835276 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835287 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835298 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835309 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835321 4715 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835332 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835345 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835357 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835371 4715 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835383 4715 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835395 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835406 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835417 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835428 4715 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835439 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835450 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835462 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835473 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835483 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835494 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835504 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835515 4715 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835526 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835536 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835547 4715 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835558 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835569 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835578 4715 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835589 4715 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835600 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835611 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835623 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835633 4715 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835644 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835687 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835698 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835710 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835720 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835731 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835741 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835754 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835775 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835786 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835797 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835810 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835822 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835832 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835844 4715 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835856 4715 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835866 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835877 4715 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835887 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835897 4715 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835927 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835939 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835949 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835960 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835970 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835982 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.835992 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836003 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836014 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836025 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836034 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836046 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836056 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836066 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836077 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836087 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836097 4715 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836109 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836119 4715 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836129 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836139 4715 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836148 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836159 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836171 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836184 4715 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836199 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836210 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836221 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836232 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836242 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836253 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836264 4715 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836274 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836285 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836295 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836306 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836317 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836329 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836339 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836349 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836360 4715 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836371 4715 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836382 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836398 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836409 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836421 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836432 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836443 4715 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836456 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836467 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836478 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836489 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836499 4715 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836510 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836521 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836533 4715 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836543 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836555 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836565 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836576 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836587 4715 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836599 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836610 4715 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836621 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836632 4715 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836652 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836663 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836674 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836685 4715 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836696 4715 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836707 4715 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836815 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.836861 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.849509 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.858686 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.867503 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 09:34:15 crc kubenswrapper[4715]: I1210 09:34:15.892038 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 09:34:15 crc kubenswrapper[4715]: W1210 09:34:15.897006 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-114382f45a28bac376226f7d4bb1bcf936fc45c291f1cb5f9e8db58b84c0c047 WatchSource:0}: Error finding container 114382f45a28bac376226f7d4bb1bcf936fc45c291f1cb5f9e8db58b84c0c047: Status 404 returned error can't find the container with id 114382f45a28bac376226f7d4bb1bcf936fc45c291f1cb5f9e8db58b84c0c047 Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.240724 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.240980 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:34:17.240941781 +0000 UTC m=+19.984488042 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.268860 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-jn8jk"] Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.269260 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-jn8jk" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.274396 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.274583 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.280221 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.285843 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.303752 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.321936 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.341517 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.341558 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.341579 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.341600 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341695 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341740 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341762 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341774 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341792 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:17.341770698 +0000 UTC m=+20.085316949 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341835 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:17.341811149 +0000 UTC m=+20.085357600 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341884 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341890 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341926 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341939 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.341929 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:17.341906802 +0000 UTC m=+20.085453053 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:16 crc kubenswrapper[4715]: E1210 09:34:16.342011 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:17.342001864 +0000 UTC m=+20.085548115 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.357248 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.400519 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.424722 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.442709 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-654kj\" (UniqueName: \"kubernetes.io/projected/1498fcdd-3a54-4845-b1bf-5f2557f6fd0d-kube-api-access-654kj\") pod \"node-resolver-jn8jk\" (UID: \"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\") " pod="openshift-dns/node-resolver-jn8jk" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.442763 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1498fcdd-3a54-4845-b1bf-5f2557f6fd0d-hosts-file\") pod \"node-resolver-jn8jk\" (UID: \"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\") " pod="openshift-dns/node-resolver-jn8jk" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.451375 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.480264 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.544052 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1498fcdd-3a54-4845-b1bf-5f2557f6fd0d-hosts-file\") pod \"node-resolver-jn8jk\" (UID: \"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\") " pod="openshift-dns/node-resolver-jn8jk" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.544821 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-654kj\" (UniqueName: \"kubernetes.io/projected/1498fcdd-3a54-4845-b1bf-5f2557f6fd0d-kube-api-access-654kj\") pod \"node-resolver-jn8jk\" (UID: \"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\") " pod="openshift-dns/node-resolver-jn8jk" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.544241 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1498fcdd-3a54-4845-b1bf-5f2557f6fd0d-hosts-file\") pod \"node-resolver-jn8jk\" (UID: \"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\") " pod="openshift-dns/node-resolver-jn8jk" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.564078 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-654kj\" (UniqueName: \"kubernetes.io/projected/1498fcdd-3a54-4845-b1bf-5f2557f6fd0d-kube-api-access-654kj\") pod \"node-resolver-jn8jk\" (UID: \"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\") " pod="openshift-dns/node-resolver-jn8jk" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.581624 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-jn8jk" Dec 10 09:34:16 crc kubenswrapper[4715]: W1210 09:34:16.593161 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1498fcdd_3a54_4845_b1bf_5f2557f6fd0d.slice/crio-201bbf519e3e4fa8456100fc63bf153035acef9404f2c202984af66f3b901426 WatchSource:0}: Error finding container 201bbf519e3e4fa8456100fc63bf153035acef9404f2c202984af66f3b901426: Status 404 returned error can't find the container with id 201bbf519e3e4fa8456100fc63bf153035acef9404f2c202984af66f3b901426 Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.622186 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.646771 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.662204 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.677305 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.683608 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.698268 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.716166 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.727977 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.742693 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.753854 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.769191 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.772597 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed"} Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.773284 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.776079 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-jn8jk" event={"ID":"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d","Type":"ContainerStarted","Data":"201bbf519e3e4fa8456100fc63bf153035acef9404f2c202984af66f3b901426"} Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.778356 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e"} Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.778381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"53c193153e80292bbbb6154fbf7ca09c361f0136a66650312bf8a92e827397b4"} Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.781391 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9"} Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.781427 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8"} Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.781454 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d5b16aeddad9ea45adaa16aee503cabfe252c83ca7f1605a711e52b4ae42d0ab"} Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.784646 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"114382f45a28bac376226f7d4bb1bcf936fc45c291f1cb5f9e8db58b84c0c047"} Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.793410 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.805838 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.820711 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.836780 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.855078 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.870743 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.884789 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.898230 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.911565 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.926957 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.937160 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.950760 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.966273 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:16 crc kubenswrapper[4715]: I1210 09:34:16.990416 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:16Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.012979 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.027423 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.045156 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.057432 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.086092 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-7pbhk"] Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.086469 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-bnh42"] Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.086650 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.087392 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.093362 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.095094 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.095095 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.095114 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.095236 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.095358 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.095542 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-98zp6"] Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.095560 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.095833 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.096459 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.096599 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.104819 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.104953 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.105282 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.120550 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.163708 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.229486 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.260755 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.264891 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265075 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-os-release\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265100 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32230eb5-c6d6-45ff-82db-78e70b283fa0-cni-binary-copy\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.265167 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:34:19.26511894 +0000 UTC m=+22.008665211 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265243 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-run-multus-certs\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265317 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-run-netns\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265355 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-hostroot\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265389 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8b26b009-53b8-4260-88a0-0f9f9553d676-mcd-auth-proxy-config\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265423 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-cnibin\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265454 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-cnibin\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265504 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8b26b009-53b8-4260-88a0-0f9f9553d676-proxy-tls\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265535 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/32230eb5-c6d6-45ff-82db-78e70b283fa0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265581 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-system-cni-dir\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265607 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8870da50-36bc-413e-96ba-89740d93d5d9-multus-daemon-config\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265644 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-multus-socket-dir-parent\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265694 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8b26b009-53b8-4260-88a0-0f9f9553d676-rootfs\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265711 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-multus-cni-dir\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265749 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8870da50-36bc-413e-96ba-89740d93d5d9-cni-binary-copy\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265815 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-var-lib-cni-bin\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265850 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265896 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-os-release\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265968 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-multus-conf-dir\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.265992 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-etc-kubernetes\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.266055 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-var-lib-kubelet\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.266077 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdtqw\" (UniqueName: \"kubernetes.io/projected/8870da50-36bc-413e-96ba-89740d93d5d9-kube-api-access-xdtqw\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.266122 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqt9k\" (UniqueName: \"kubernetes.io/projected/32230eb5-c6d6-45ff-82db-78e70b283fa0-kube-api-access-hqt9k\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.266150 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-var-lib-cni-multus\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.266197 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-run-k8s-cni-cncf-io\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.266229 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98m8k\" (UniqueName: \"kubernetes.io/projected/8b26b009-53b8-4260-88a0-0f9f9553d676-kube-api-access-98m8k\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.266274 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-system-cni-dir\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.304004 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.348759 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367238 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-os-release\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367276 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367294 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367335 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367352 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367368 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-multus-conf-dir\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367403 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-etc-kubernetes\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367419 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-var-lib-cni-multus\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367425 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-os-release\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367468 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-multus-conf-dir\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367434 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-var-lib-kubelet\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.367478 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367507 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdtqw\" (UniqueName: \"kubernetes.io/projected/8870da50-36bc-413e-96ba-89740d93d5d9-kube-api-access-xdtqw\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367525 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqt9k\" (UniqueName: \"kubernetes.io/projected/32230eb5-c6d6-45ff-82db-78e70b283fa0-kube-api-access-hqt9k\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367527 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-var-lib-kubelet\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.367526 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367507 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-var-lib-cni-multus\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.367569 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367619 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-etc-kubernetes\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.367642 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.367535 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367800 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.367573 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:19.367546164 +0000 UTC m=+22.111092415 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.367871 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98m8k\" (UniqueName: \"kubernetes.io/projected/8b26b009-53b8-4260-88a0-0f9f9553d676-kube-api-access-98m8k\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.367927 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:19.367887594 +0000 UTC m=+22.111434005 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.367962 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:19.367951205 +0000 UTC m=+22.111497656 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368004 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-system-cni-dir\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368039 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-run-k8s-cni-cncf-io\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368070 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-os-release\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368096 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32230eb5-c6d6-45ff-82db-78e70b283fa0-cni-binary-copy\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368116 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-run-multus-certs\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368133 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-system-cni-dir\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368143 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8b26b009-53b8-4260-88a0-0f9f9553d676-mcd-auth-proxy-config\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368193 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-run-k8s-cni-cncf-io\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368210 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-run-multus-certs\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368205 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-os-release\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368235 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-cnibin\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368275 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-run-netns\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368286 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-cnibin\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368296 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-hostroot\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368328 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368374 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-hostroot\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368390 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-cnibin\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368421 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-cnibin\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368433 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8b26b009-53b8-4260-88a0-0f9f9553d676-proxy-tls\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368332 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-run-netns\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368456 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/32230eb5-c6d6-45ff-82db-78e70b283fa0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368480 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-system-cni-dir\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368498 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8870da50-36bc-413e-96ba-89740d93d5d9-multus-daemon-config\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368514 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-multus-socket-dir-parent\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368534 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8b26b009-53b8-4260-88a0-0f9f9553d676-rootfs\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368540 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32230eb5-c6d6-45ff-82db-78e70b283fa0-system-cni-dir\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368561 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-multus-cni-dir\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368579 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8870da50-36bc-413e-96ba-89740d93d5d9-cni-binary-copy\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368600 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-var-lib-cni-bin\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.368599 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.368659 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.368675 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368598 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8b26b009-53b8-4260-88a0-0f9f9553d676-rootfs\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368633 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-host-var-lib-cni-bin\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.368704 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:19.368695557 +0000 UTC m=+22.112241798 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368723 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-multus-socket-dir-parent\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.368802 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8870da50-36bc-413e-96ba-89740d93d5d9-multus-cni-dir\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.369079 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32230eb5-c6d6-45ff-82db-78e70b283fa0-cni-binary-copy\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.369131 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8b26b009-53b8-4260-88a0-0f9f9553d676-mcd-auth-proxy-config\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.369170 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/32230eb5-c6d6-45ff-82db-78e70b283fa0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.369442 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8870da50-36bc-413e-96ba-89740d93d5d9-multus-daemon-config\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.369528 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8870da50-36bc-413e-96ba-89740d93d5d9-cni-binary-copy\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.373406 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8b26b009-53b8-4260-88a0-0f9f9553d676-proxy-tls\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.400183 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98m8k\" (UniqueName: \"kubernetes.io/projected/8b26b009-53b8-4260-88a0-0f9f9553d676-kube-api-access-98m8k\") pod \"machine-config-daemon-7pbhk\" (UID: \"8b26b009-53b8-4260-88a0-0f9f9553d676\") " pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.403996 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.426288 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.426962 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdtqw\" (UniqueName: \"kubernetes.io/projected/8870da50-36bc-413e-96ba-89740d93d5d9-kube-api-access-xdtqw\") pod \"multus-98zp6\" (UID: \"8870da50-36bc-413e-96ba-89740d93d5d9\") " pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.427237 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqt9k\" (UniqueName: \"kubernetes.io/projected/32230eb5-c6d6-45ff-82db-78e70b283fa0-kube-api-access-hqt9k\") pod \"multus-additional-cni-plugins-bnh42\" (UID: \"32230eb5-c6d6-45ff-82db-78e70b283fa0\") " pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.428704 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-98zp6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.431656 4715 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.431971 4715 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432468 4715 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432516 4715 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432540 4715 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432557 4715 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432576 4715 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432680 4715 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432703 4715 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432712 4715 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432737 4715 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432746 4715 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.432829 4715 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.432843 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/events/networking-console-plugin-85b44fc459-gdk6g.187fd0e3cbd43256\": read tcp 38.102.83.201:43510->38.102.83.201:6443: use of closed network connection" event="&Event{ObjectMeta:{networking-console-plugin-85b44fc459-gdk6g.187fd0e3cbd43256 openshift-network-console 26510 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-network-console,Name:networking-console-plugin-85b44fc459-gdk6g,UID:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8,APIVersion:v1,ResourceVersion:25349,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"networking-console-plugin-cert\" : object \"openshift-network-console\"/\"networking-console-plugin-cert\" not registered,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 09:34:15 +0000 UTC,LastTimestamp:2025-12-10 09:34:17.367827722 +0000 UTC m=+20.111373973,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.553385 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vspkd"] Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.554176 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.557264 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.557310 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.557392 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.557567 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.557685 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.557695 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.557991 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572328 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-ovn\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572379 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-log-socket\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572401 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-netd\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572449 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-script-lib\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572476 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-etc-openvswitch\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572497 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-var-lib-openvswitch\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572518 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572542 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-env-overrides\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572570 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-kubelet\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572591 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-systemd-units\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572613 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-systemd\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572642 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-node-log\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572666 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbxh6\" (UniqueName: \"kubernetes.io/projected/2a44160c-cf30-4b13-b82c-4c402d967dd8-kube-api-access-sbxh6\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572686 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-netns\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572706 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-ovn-kubernetes\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572726 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-bin\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572749 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-slash\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572770 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-config\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572791 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovn-node-metrics-cert\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.572811 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-openvswitch\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.604486 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.604607 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.604900 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.604988 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.605079 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:17 crc kubenswrapper[4715]: E1210 09:34:17.605180 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.609048 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.609774 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.610954 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.611690 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.612758 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.613249 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.613781 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.614816 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.615971 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.617123 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.617818 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.619140 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.619674 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.620246 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.621108 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.621612 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.622513 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.622864 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.623500 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.624565 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.625062 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.626041 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.626484 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.627625 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.628901 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.629601 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.630745 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.631275 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.632425 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.632873 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.633745 4715 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.633861 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.635681 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.636615 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.637078 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.638579 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.639338 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.640267 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.640890 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.642234 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.642735 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.643881 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.644570 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.645569 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.646126 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.647056 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.647589 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.648676 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.649175 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.650034 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.650485 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.651490 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.652087 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.652571 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674628 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-systemd\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674680 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-node-log\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674699 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbxh6\" (UniqueName: \"kubernetes.io/projected/2a44160c-cf30-4b13-b82c-4c402d967dd8-kube-api-access-sbxh6\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674716 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-netns\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674732 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-ovn-kubernetes\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674747 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-bin\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674763 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-slash\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674781 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-config\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674796 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovn-node-metrics-cert\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674812 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-openvswitch\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674829 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-ovn\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674843 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-log-socket\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674860 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-netd\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674859 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-slash\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674896 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-script-lib\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-log-socket\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674832 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-netns\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674941 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-bin\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674956 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-ovn\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674970 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-openvswitch\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.674846 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-ovn-kubernetes\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675006 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-etc-openvswitch\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675018 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-netd\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675093 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-etc-openvswitch\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675120 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-var-lib-openvswitch\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675154 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675172 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-env-overrides\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675195 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-var-lib-openvswitch\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675217 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675304 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-kubelet\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675323 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-systemd-units\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675378 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-systemd-units\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675404 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-kubelet\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675569 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-env-overrides\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675738 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-config\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675817 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-node-log\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.675878 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-systemd\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.676039 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-script-lib\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.680438 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovn-node-metrics-cert\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.698422 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbxh6\" (UniqueName: \"kubernetes.io/projected/2a44160c-cf30-4b13-b82c-4c402d967dd8-kube-api-access-sbxh6\") pod \"ovnkube-node-vspkd\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.716380 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-bnh42" Dec 10 09:34:17 crc kubenswrapper[4715]: W1210 09:34:17.727646 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32230eb5_c6d6_45ff_82db_78e70b283fa0.slice/crio-ac0c4906b6ffeaba72da4d0c809ecce107c38e842364f982f8c5ee554ef8f214 WatchSource:0}: Error finding container ac0c4906b6ffeaba72da4d0c809ecce107c38e842364f982f8c5ee554ef8f214: Status 404 returned error can't find the container with id ac0c4906b6ffeaba72da4d0c809ecce107c38e842364f982f8c5ee554ef8f214 Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.791647 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9"} Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.791695 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5"} Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.791705 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"6fb3fa86d7f3a52225466795e7f9545e98bf0a9efe559da6e74d86cdf7bf854b"} Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.823214 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-98zp6" event={"ID":"8870da50-36bc-413e-96ba-89740d93d5d9","Type":"ContainerStarted","Data":"09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855"} Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.823276 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-98zp6" event={"ID":"8870da50-36bc-413e-96ba-89740d93d5d9","Type":"ContainerStarted","Data":"4556781107aa4a36d9d8df2ed040620550bc843e6bca39738c772639d41b0220"} Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.837439 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-jn8jk" event={"ID":"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d","Type":"ContainerStarted","Data":"485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de"} Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.854779 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" event={"ID":"32230eb5-c6d6-45ff-82db-78e70b283fa0","Type":"ContainerStarted","Data":"ac0c4906b6ffeaba72da4d0c809ecce107c38e842364f982f8c5ee554ef8f214"} Dec 10 09:34:17 crc kubenswrapper[4715]: I1210 09:34:17.867282 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:18 crc kubenswrapper[4715]: W1210 09:34:18.141074 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a44160c_cf30_4b13_b82c_4c402d967dd8.slice/crio-9761e111d4530a07bf0cdc6c800b0cd4dd70e66cd1972fbd1ef86961b3879cf0 WatchSource:0}: Error finding container 9761e111d4530a07bf0cdc6c800b0cd4dd70e66cd1972fbd1ef86961b3879cf0: Status 404 returned error can't find the container with id 9761e111d4530a07bf0cdc6c800b0cd4dd70e66cd1972fbd1ef86961b3879cf0 Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.249223 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.296604 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.316443 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.318217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.318251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.318269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.318390 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.325767 4715 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.326105 4715 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.327256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.327287 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.327300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.327316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.327328 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: E1210 09:34:18.343607 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.347229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.347254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.347262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.347275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.347286 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: E1210 09:34:18.357141 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.359932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.359961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.359969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.359981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.359989 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: E1210 09:34:18.370873 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.374717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.374964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.375064 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.375171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.375270 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: E1210 09:34:18.386716 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.390476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.390643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.390738 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.390829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.390977 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: E1210 09:34:18.402699 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: E1210 09:34:18.403246 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.410356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.410518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.410577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.410651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.410711 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.412298 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.449755 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.465073 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.476327 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.497416 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.507623 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.513373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.513415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.513427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.513445 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.513458 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.565525 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.573203 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.588631 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.609883 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.615507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.615547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.615560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.615576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.615587 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.627422 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.627936 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.644587 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.657642 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.664071 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.674115 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.695744 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.708966 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.719794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.719841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.719852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.719867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.719878 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.723644 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.793163 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.807289 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.819369 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.819375 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.822103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.822134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.822145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.822161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.822173 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.833647 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:18Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.858646 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" event={"ID":"32230eb5-c6d6-45ff-82db-78e70b283fa0","Type":"ContainerStarted","Data":"49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c"} Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.860104 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f"} Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.861155 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06" exitCode=0 Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.861180 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06"} Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.861195 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"9761e111d4530a07bf0cdc6c800b0cd4dd70e66cd1972fbd1ef86961b3879cf0"} Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.923761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.923794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.923806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.923822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.923834 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:18Z","lastTransitionTime":"2025-12-10T09:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.944958 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.958315 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 09:34:18 crc kubenswrapper[4715]: I1210 09:34:18.960345 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.016960 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.025969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.025995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.026004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.026018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.026027 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:19Z","lastTransitionTime":"2025-12-10T09:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.026807 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.038429 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.060728 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.080010 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.082562 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.092575 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.094074 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.101415 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.114103 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.125199 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.130752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.130784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.130795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.130831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.130846 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:19Z","lastTransitionTime":"2025-12-10T09:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.140658 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.154086 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.169419 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.187754 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.209224 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.233692 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.233891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.233995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.234064 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.234124 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:19Z","lastTransitionTime":"2025-12-10T09:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.248562 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.261609 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.272727 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.292878 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.295002 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.295545 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:34:23.295522147 +0000 UTC m=+26.039068398 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.306254 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.326291 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.336610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.336671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.336688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.336712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.336727 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:19Z","lastTransitionTime":"2025-12-10T09:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.351028 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.367749 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.380764 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.392354 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.395623 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.395668 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.395690 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.395708 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.395819 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.395863 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:23.395848199 +0000 UTC m=+26.139394450 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.396142 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.396158 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.396169 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.396196 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:23.396188069 +0000 UTC m=+26.139734320 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.396248 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.396257 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.396265 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.396287 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:23.396280242 +0000 UTC m=+26.139826493 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.396317 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.396339 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:23.396333603 +0000 UTC m=+26.139879854 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.404890 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.416065 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.433175 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.439154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.439182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.439190 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.439208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.439217 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:19Z","lastTransitionTime":"2025-12-10T09:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.541122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.541166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.541180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.541197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.541211 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:19Z","lastTransitionTime":"2025-12-10T09:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.603850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.603887 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.603860 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.604016 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.604109 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:19 crc kubenswrapper[4715]: E1210 09:34:19.604179 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.643947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.643982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.643991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.644006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.644016 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:19Z","lastTransitionTime":"2025-12-10T09:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.747580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.748004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.748016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.748037 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.748052 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:19Z","lastTransitionTime":"2025-12-10T09:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.821001 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zddqm"] Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.821508 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.823287 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.823315 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.823396 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.823718 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.837206 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.848647 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.850487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.850559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.850572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.850597 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.850616 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:19Z","lastTransitionTime":"2025-12-10T09:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.865270 4715 generic.go:334] "Generic (PLEG): container finished" podID="32230eb5-c6d6-45ff-82db-78e70b283fa0" containerID="49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c" exitCode=0 Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.865346 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" event={"ID":"32230eb5-c6d6-45ff-82db-78e70b283fa0","Type":"ContainerDied","Data":"49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.868362 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.869086 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.869951 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178" exitCode=1 Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.870004 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.870038 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.870053 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.870067 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.870080 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.891003 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.900676 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/57659947-16fe-444e-a0d9-bc825e0aee39-serviceca\") pod \"node-ca-zddqm\" (UID: \"57659947-16fe-444e-a0d9-bc825e0aee39\") " pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.900713 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57659947-16fe-444e-a0d9-bc825e0aee39-host\") pod \"node-ca-zddqm\" (UID: \"57659947-16fe-444e-a0d9-bc825e0aee39\") " pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.900739 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4grj4\" (UniqueName: \"kubernetes.io/projected/57659947-16fe-444e-a0d9-bc825e0aee39-kube-api-access-4grj4\") pod \"node-ca-zddqm\" (UID: \"57659947-16fe-444e-a0d9-bc825e0aee39\") " pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.907181 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.919454 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.929775 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.945372 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.960029 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.960518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.960540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.960554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.960572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.960585 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:19Z","lastTransitionTime":"2025-12-10T09:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.977993 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:19 crc kubenswrapper[4715]: I1210 09:34:19.994695 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.002679 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/57659947-16fe-444e-a0d9-bc825e0aee39-serviceca\") pod \"node-ca-zddqm\" (UID: \"57659947-16fe-444e-a0d9-bc825e0aee39\") " pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.003061 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57659947-16fe-444e-a0d9-bc825e0aee39-host\") pod \"node-ca-zddqm\" (UID: \"57659947-16fe-444e-a0d9-bc825e0aee39\") " pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.003235 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4grj4\" (UniqueName: \"kubernetes.io/projected/57659947-16fe-444e-a0d9-bc825e0aee39-kube-api-access-4grj4\") pod \"node-ca-zddqm\" (UID: \"57659947-16fe-444e-a0d9-bc825e0aee39\") " pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.003783 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57659947-16fe-444e-a0d9-bc825e0aee39-host\") pod \"node-ca-zddqm\" (UID: \"57659947-16fe-444e-a0d9-bc825e0aee39\") " pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.003897 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/57659947-16fe-444e-a0d9-bc825e0aee39-serviceca\") pod \"node-ca-zddqm\" (UID: \"57659947-16fe-444e-a0d9-bc825e0aee39\") " pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.021232 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.026546 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4grj4\" (UniqueName: \"kubernetes.io/projected/57659947-16fe-444e-a0d9-bc825e0aee39-kube-api-access-4grj4\") pod \"node-ca-zddqm\" (UID: \"57659947-16fe-444e-a0d9-bc825e0aee39\") " pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.035472 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.048077 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.060720 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.063179 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.063209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.063217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.063232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.063242 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:20Z","lastTransitionTime":"2025-12-10T09:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.072550 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.089508 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.104856 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.118121 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.132201 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.151302 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.154591 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zddqm" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.165127 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: W1210 09:34:20.179125 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57659947_16fe_444e_a0d9_bc825e0aee39.slice/crio-3e6ecb8028c025ca9755f1106ed08b90d595de40d9f75f097a79ad547e8d527e WatchSource:0}: Error finding container 3e6ecb8028c025ca9755f1106ed08b90d595de40d9f75f097a79ad547e8d527e: Status 404 returned error can't find the container with id 3e6ecb8028c025ca9755f1106ed08b90d595de40d9f75f097a79ad547e8d527e Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.179868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.179906 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.179933 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.179952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.179965 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:20Z","lastTransitionTime":"2025-12-10T09:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.191579 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.217104 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.230275 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.245758 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.320479 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.323025 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.323081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.323096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.323119 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.323136 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:20Z","lastTransitionTime":"2025-12-10T09:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.340423 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.353895 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.372320 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.432106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.432130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.432138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.432150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.432159 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:20Z","lastTransitionTime":"2025-12-10T09:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.534388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.534426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.534437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.534452 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.534464 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:20Z","lastTransitionTime":"2025-12-10T09:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.660071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.660111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.660126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.660147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.660162 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:20Z","lastTransitionTime":"2025-12-10T09:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.762840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.762882 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.762893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.762920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.762931 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:20Z","lastTransitionTime":"2025-12-10T09:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.865403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.865443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.865453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.865469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.865481 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:20Z","lastTransitionTime":"2025-12-10T09:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.874796 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" event={"ID":"32230eb5-c6d6-45ff-82db-78e70b283fa0","Type":"ContainerStarted","Data":"462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.878000 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.878978 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.880611 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zddqm" event={"ID":"57659947-16fe-444e-a0d9-bc825e0aee39","Type":"ContainerStarted","Data":"dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.880677 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zddqm" event={"ID":"57659947-16fe-444e-a0d9-bc825e0aee39","Type":"ContainerStarted","Data":"3e6ecb8028c025ca9755f1106ed08b90d595de40d9f75f097a79ad547e8d527e"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.892529 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.915439 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.927241 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.937618 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.951140 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.964745 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.967579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.967611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.967619 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.967635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.967644 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:20Z","lastTransitionTime":"2025-12-10T09:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.981640 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:20 crc kubenswrapper[4715]: I1210 09:34:20.996889 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.023214 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.037291 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.050205 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.063872 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.070296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.070338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.070347 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.070361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.070371 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:21Z","lastTransitionTime":"2025-12-10T09:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.076041 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.086879 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.112336 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.131754 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.143232 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.155094 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.168984 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.172982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.173107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.173214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.173322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.173448 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:21Z","lastTransitionTime":"2025-12-10T09:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.182765 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.199952 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.217783 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.232746 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.245149 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.257327 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.269281 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.276044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.276222 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.276309 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.276406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.276483 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:21Z","lastTransitionTime":"2025-12-10T09:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.283036 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.295649 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.308763 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.319135 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.379279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.379311 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.379320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.379334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.379344 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:21Z","lastTransitionTime":"2025-12-10T09:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.482041 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.482080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.482091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.482106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.482116 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:21Z","lastTransitionTime":"2025-12-10T09:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.552501 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.556240 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.567296 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.580906 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.584406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.584436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.584445 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.584458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.584466 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:21Z","lastTransitionTime":"2025-12-10T09:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.591702 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.602261 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.604418 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.604445 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.604499 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:21 crc kubenswrapper[4715]: E1210 09:34:21.604566 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:21 crc kubenswrapper[4715]: E1210 09:34:21.604675 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:21 crc kubenswrapper[4715]: E1210 09:34:21.604770 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.619580 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.632488 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.646697 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.658234 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.680086 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.687677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.687725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.687736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.687759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.687774 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:21Z","lastTransitionTime":"2025-12-10T09:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.696624 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.707687 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.730958 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.773643 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.790363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.790434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.790461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.790492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.790517 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:21Z","lastTransitionTime":"2025-12-10T09:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.811157 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.855202 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.891155 4715 generic.go:334] "Generic (PLEG): container finished" podID="32230eb5-c6d6-45ff-82db-78e70b283fa0" containerID="462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741" exitCode=0 Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.891243 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" event={"ID":"32230eb5-c6d6-45ff-82db-78e70b283fa0","Type":"ContainerDied","Data":"462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741"} Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.891758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.891803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.891814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.891828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.891837 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:21Z","lastTransitionTime":"2025-12-10T09:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.896217 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.928065 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.970026 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:21Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.994462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.994688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.994767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.994879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:21 crc kubenswrapper[4715]: I1210 09:34:21.994978 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:21Z","lastTransitionTime":"2025-12-10T09:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.010281 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.046953 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.088871 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.098461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.098491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.098500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.098513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.098522 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:22Z","lastTransitionTime":"2025-12-10T09:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.136365 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.217752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.217794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.217802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.217822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.217830 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:22Z","lastTransitionTime":"2025-12-10T09:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.226584 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.241990 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.348036 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.352199 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.352234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.352244 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.352260 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.352270 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:22Z","lastTransitionTime":"2025-12-10T09:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.371396 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.386687 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.400376 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.427173 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.449237 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.460680 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.460718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.460728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.460744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.460758 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:22Z","lastTransitionTime":"2025-12-10T09:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.489265 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.529717 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.562774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.562824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.562837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.562904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.562919 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:22Z","lastTransitionTime":"2025-12-10T09:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.568618 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.607368 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.649617 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.665504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.665538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.665549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.665564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.665575 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:22Z","lastTransitionTime":"2025-12-10T09:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.687069 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.733626 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.767871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.767899 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.767907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.767922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.767930 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:22Z","lastTransitionTime":"2025-12-10T09:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.778415 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.808168 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.852338 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.870236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.870262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.870272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.870289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.870298 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:22Z","lastTransitionTime":"2025-12-10T09:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.895014 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.897691 4715 generic.go:334] "Generic (PLEG): container finished" podID="32230eb5-c6d6-45ff-82db-78e70b283fa0" containerID="740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b" exitCode=0 Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.897764 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" event={"ID":"32230eb5-c6d6-45ff-82db-78e70b283fa0","Type":"ContainerDied","Data":"740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.902422 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.903312 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.936431 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.972251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.972306 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.972319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.972343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.972359 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:22Z","lastTransitionTime":"2025-12-10T09:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:22 crc kubenswrapper[4715]: I1210 09:34:22.979408 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:22Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.010310 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.052827 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.076016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.076069 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.076082 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.076102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.076114 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:23Z","lastTransitionTime":"2025-12-10T09:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.098291 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.129432 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.167660 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.178872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.178900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.178909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.178947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.178957 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:23Z","lastTransitionTime":"2025-12-10T09:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.209400 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.251466 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.281703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.281739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.281750 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.281766 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.281778 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:23Z","lastTransitionTime":"2025-12-10T09:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.289360 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.343855 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.356808 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.357055 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:34:31.357020271 +0000 UTC m=+34.100566532 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.372815 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.383648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.383699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.383710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.383729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.383741 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:23Z","lastTransitionTime":"2025-12-10T09:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.409896 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.452684 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.458256 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.458317 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.458346 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.458376 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458475 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458496 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458525 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458545 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458578 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458526 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:31.458510427 +0000 UTC m=+34.202056688 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458584 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458616 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:31.45859693 +0000 UTC m=+34.202143201 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458625 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458640 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:31.458628211 +0000 UTC m=+34.202174482 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458641 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.458744 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:31.458719484 +0000 UTC m=+34.202265785 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.485416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.485475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.485492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.485514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.485550 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:23Z","lastTransitionTime":"2025-12-10T09:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.488891 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.530472 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.569394 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.588385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.588432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.588442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.588459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.588470 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:23Z","lastTransitionTime":"2025-12-10T09:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.604723 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.604776 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.604736 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.604960 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.604977 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:23 crc kubenswrapper[4715]: E1210 09:34:23.605040 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.608599 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.647418 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.690086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.690134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.690143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.690157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.690167 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:23Z","lastTransitionTime":"2025-12-10T09:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.792905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.792957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.792966 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.792981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.792994 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:23Z","lastTransitionTime":"2025-12-10T09:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.895625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.895692 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.895724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.895745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.895784 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:23Z","lastTransitionTime":"2025-12-10T09:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.909747 4715 generic.go:334] "Generic (PLEG): container finished" podID="32230eb5-c6d6-45ff-82db-78e70b283fa0" containerID="a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0" exitCode=0 Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.909798 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" event={"ID":"32230eb5-c6d6-45ff-82db-78e70b283fa0","Type":"ContainerDied","Data":"a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0"} Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.922157 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.935404 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.946683 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.960299 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.972824 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:23 crc kubenswrapper[4715]: I1210 09:34:23.988081 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.000037 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.000450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.000467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.000489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.000499 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:24Z","lastTransitionTime":"2025-12-10T09:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.001348 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:23Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.017360 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.039820 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.054331 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.089987 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.102828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.102856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.102864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.102878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.102886 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:24Z","lastTransitionTime":"2025-12-10T09:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.136391 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.174113 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.207160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.207216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.207229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.207253 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.207268 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:24Z","lastTransitionTime":"2025-12-10T09:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.214356 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.250486 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.309656 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.309713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.309725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.309746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.309758 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:24Z","lastTransitionTime":"2025-12-10T09:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.413785 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.413848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.413867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.413891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.413916 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:24Z","lastTransitionTime":"2025-12-10T09:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.518212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.518245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.518253 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.518265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.518274 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:24Z","lastTransitionTime":"2025-12-10T09:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.621614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.621676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.621695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.621721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.621741 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:24Z","lastTransitionTime":"2025-12-10T09:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.723407 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.723876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.723962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.724050 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.724132 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:24Z","lastTransitionTime":"2025-12-10T09:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.826674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.826713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.826726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.826748 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.826762 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:24Z","lastTransitionTime":"2025-12-10T09:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.917241 4715 generic.go:334] "Generic (PLEG): container finished" podID="32230eb5-c6d6-45ff-82db-78e70b283fa0" containerID="5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854" exitCode=0 Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.917314 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" event={"ID":"32230eb5-c6d6-45ff-82db-78e70b283fa0","Type":"ContainerDied","Data":"5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.923726 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.924848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.925226 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.925274 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.925372 4715 scope.go:117] "RemoveContainer" containerID="e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.930135 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.930169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.930221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.930243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.930258 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:24Z","lastTransitionTime":"2025-12-10T09:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.931368 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.950902 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.971977 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.984964 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.985402 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:24 crc kubenswrapper[4715]: I1210 09:34:24.997029 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:24Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.010420 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.023585 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.039219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.039262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.039275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.039293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.039365 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:25Z","lastTransitionTime":"2025-12-10T09:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.040141 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.053629 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.069243 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.081434 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.085902 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.099341 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.109860 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.121035 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.156936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.157050 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.157065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.157086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.157102 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:25Z","lastTransitionTime":"2025-12-10T09:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.188333 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.208160 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.229635 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.245108 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.255635 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.259304 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.259345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.259353 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.259368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.259378 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:25Z","lastTransitionTime":"2025-12-10T09:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.268851 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.280672 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.289948 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.299450 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.316553 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.327631 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.338430 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.351096 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.361037 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.361070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.361082 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.361100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.361111 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:25Z","lastTransitionTime":"2025-12-10T09:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.369105 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.408130 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.459644 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-acl-logging ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-acl-logging ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.463112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.463154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.463166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.463182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.463192 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:25Z","lastTransitionTime":"2025-12-10T09:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.566249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.566296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.566312 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.566329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.566341 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:25Z","lastTransitionTime":"2025-12-10T09:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.604157 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.604223 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:25 crc kubenswrapper[4715]: E1210 09:34:25.604312 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.604365 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:25 crc kubenswrapper[4715]: E1210 09:34:25.604472 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:25 crc kubenswrapper[4715]: E1210 09:34:25.604544 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.667921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.667965 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.667973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.667988 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.667997 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:25Z","lastTransitionTime":"2025-12-10T09:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.770832 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.770866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.770878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.770894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.770905 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:25Z","lastTransitionTime":"2025-12-10T09:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.873255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.873323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.873347 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.873382 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.873407 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:25Z","lastTransitionTime":"2025-12-10T09:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.932555 4715 generic.go:334] "Generic (PLEG): container finished" podID="32230eb5-c6d6-45ff-82db-78e70b283fa0" containerID="918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501" exitCode=0 Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.932635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" event={"ID":"32230eb5-c6d6-45ff-82db-78e70b283fa0","Type":"ContainerDied","Data":"918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.945065 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.946481 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.946928 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.952667 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.970857 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.975799 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.975852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.975867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.975887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.975900 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:25Z","lastTransitionTime":"2025-12-10T09:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:25 crc kubenswrapper[4715]: I1210 09:34:25.995429 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:25Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.011343 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.024638 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.038980 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.060459 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.076533 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.081123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.081155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.081169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.081187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.081200 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:26Z","lastTransitionTime":"2025-12-10T09:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.090630 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.106948 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.121270 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.135736 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.158346 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-acl-logging ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-acl-logging ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.172422 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.187789 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.190159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.190186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.190195 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.190208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.190217 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:26Z","lastTransitionTime":"2025-12-10T09:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.208028 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.223225 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.237721 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.249988 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.267081 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.292761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.293365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.293440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.293542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.293704 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:26Z","lastTransitionTime":"2025-12-10T09:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.292784 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.329631 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.370056 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.397401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.397732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.397853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.397991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.398157 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:26Z","lastTransitionTime":"2025-12-10T09:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.413019 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.454466 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.489200 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.502883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.502922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.502931 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.502959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.502973 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:26Z","lastTransitionTime":"2025-12-10T09:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.529321 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.574453 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.606171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.606209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.606218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.606233 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.606241 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:26Z","lastTransitionTime":"2025-12-10T09:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.622447 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.649836 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.709465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.709508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.709518 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.709537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.709550 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:26Z","lastTransitionTime":"2025-12-10T09:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.813126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.813182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.813198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.813219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.813230 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:26Z","lastTransitionTime":"2025-12-10T09:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.915454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.915507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.915520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.915538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.915565 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:26Z","lastTransitionTime":"2025-12-10T09:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.953614 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.954867 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" event={"ID":"32230eb5-c6d6-45ff-82db-78e70b283fa0","Type":"ContainerStarted","Data":"070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807"} Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.967368 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.978881 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:26 crc kubenswrapper[4715]: I1210 09:34:26.995420 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:26Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.007017 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.018455 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.018521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.018545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.018569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.018583 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:27Z","lastTransitionTime":"2025-12-10T09:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.019792 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.032017 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.045740 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.058658 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.068074 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.078902 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.094062 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.121500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.121543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.121554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.121573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.121587 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:27Z","lastTransitionTime":"2025-12-10T09:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.129414 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.169203 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.210289 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.223554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.223621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.223634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.223651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.223661 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:27Z","lastTransitionTime":"2025-12-10T09:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.254824 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.326129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.326169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.326180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.326197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.326209 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:27Z","lastTransitionTime":"2025-12-10T09:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.428828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.428860 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.428867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.428880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.428889 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:27Z","lastTransitionTime":"2025-12-10T09:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.531679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.531716 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.531724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.531741 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.531749 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:27Z","lastTransitionTime":"2025-12-10T09:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.604737 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.604737 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:27 crc kubenswrapper[4715]: E1210 09:34:27.604883 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.604840 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:27 crc kubenswrapper[4715]: E1210 09:34:27.605121 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:27 crc kubenswrapper[4715]: E1210 09:34:27.605227 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.630421 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.634587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.634628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.634637 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.634655 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.634666 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:27Z","lastTransitionTime":"2025-12-10T09:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.644357 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.660068 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.679791 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.703241 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.730967 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.736134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.736161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.736170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.736186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.736197 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:27Z","lastTransitionTime":"2025-12-10T09:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.752549 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.779618 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.799493 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.811821 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.827615 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.838770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.838970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.839065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.839157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.839243 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:27Z","lastTransitionTime":"2025-12-10T09:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.915733 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.917456 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.991755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.991815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.991840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.991875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.991902 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:27Z","lastTransitionTime":"2025-12-10T09:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:27 crc kubenswrapper[4715]: I1210 09:34:27.997297 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.018159 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.034255 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.046992 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.061147 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.083046 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.094574 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.094616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.094631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.094649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.094664 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.096481 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.111321 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.136957 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.170337 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.188133 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.196999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.197241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.197356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.197463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.197572 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.207941 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.248191 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.290577 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.300717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.300757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.300768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.300787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.300799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.329302 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.367945 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.403841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.403932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.403945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.403967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.403980 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.409745 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.450863 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.506174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.506204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.506214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.506231 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.506244 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.555875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.555976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.556209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.556254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.556269 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: E1210 09:34:28.569523 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.572914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.572959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.572971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.572990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.573000 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: E1210 09:34:28.585269 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.588492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.588524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.588532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.588547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.588558 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: E1210 09:34:28.598347 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.601146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.601172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.601181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.601196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.601205 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: E1210 09:34:28.610800 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.613525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.613562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.613575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.613591 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.613601 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: E1210 09:34:28.623168 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:28Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:28 crc kubenswrapper[4715]: E1210 09:34:28.623294 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.624807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.624833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.624841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.624858 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.624868 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.727000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.727042 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.727056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.727073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.727084 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.829027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.829066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.829075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.829094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.829104 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.930753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.930792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.930802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.930814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:28 crc kubenswrapper[4715]: I1210 09:34:28.930826 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:28Z","lastTransitionTime":"2025-12-10T09:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.033314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.033361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.033381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.033399 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.033410 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:29Z","lastTransitionTime":"2025-12-10T09:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.094894 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27"] Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.095403 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: W1210 09:34:29.101297 4715 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert": failed to list *v1.Secret: secrets "ovn-control-plane-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 10 09:34:29 crc kubenswrapper[4715]: E1210 09:34:29.101367 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-control-plane-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 09:34:29 crc kubenswrapper[4715]: W1210 09:34:29.101324 4715 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd": failed to list *v1.Secret: secrets "ovn-kubernetes-control-plane-dockercfg-gs7dd" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 10 09:34:29 crc kubenswrapper[4715]: E1210 09:34:29.101409 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-gs7dd\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-control-plane-dockercfg-gs7dd\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.112607 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.125089 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc3c82fe-6594-41d0-96d5-d8977ac76b15-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.125123 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hclwf\" (UniqueName: \"kubernetes.io/projected/cc3c82fe-6594-41d0-96d5-d8977ac76b15-kube-api-access-hclwf\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.125148 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc3c82fe-6594-41d0-96d5-d8977ac76b15-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.125164 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc3c82fe-6594-41d0-96d5-d8977ac76b15-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.129461 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.135606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.135638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.135646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.135659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.135668 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:29Z","lastTransitionTime":"2025-12-10T09:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.143994 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.160525 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.176055 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.191762 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.204035 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.217480 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.225976 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc3c82fe-6594-41d0-96d5-d8977ac76b15-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.226036 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hclwf\" (UniqueName: \"kubernetes.io/projected/cc3c82fe-6594-41d0-96d5-d8977ac76b15-kube-api-access-hclwf\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.226068 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc3c82fe-6594-41d0-96d5-d8977ac76b15-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.226103 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc3c82fe-6594-41d0-96d5-d8977ac76b15-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.226587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cc3c82fe-6594-41d0-96d5-d8977ac76b15-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.226835 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cc3c82fe-6594-41d0-96d5-d8977ac76b15-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.237587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.237639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.237648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.237665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.237675 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:29Z","lastTransitionTime":"2025-12-10T09:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.241879 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.253706 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hclwf\" (UniqueName: \"kubernetes.io/projected/cc3c82fe-6594-41d0-96d5-d8977ac76b15-kube-api-access-hclwf\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.256502 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.271390 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.286588 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.297749 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.308013 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.317415 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.333513 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:29Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.340378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.340409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.340420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.340435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.340445 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:29Z","lastTransitionTime":"2025-12-10T09:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.442779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.442842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.442858 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.442883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.442902 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:29Z","lastTransitionTime":"2025-12-10T09:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.545680 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.545736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.545753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.545777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.545794 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:29Z","lastTransitionTime":"2025-12-10T09:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.604189 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.604562 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:29 crc kubenswrapper[4715]: E1210 09:34:29.604700 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.604766 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:29 crc kubenswrapper[4715]: E1210 09:34:29.604830 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:29 crc kubenswrapper[4715]: E1210 09:34:29.605043 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.647622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.647871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.647959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.648033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.648114 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:29Z","lastTransitionTime":"2025-12-10T09:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.751173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.751219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.751227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.751240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.751249 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:29Z","lastTransitionTime":"2025-12-10T09:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.854718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.854819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.854841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.854872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.854897 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:29Z","lastTransitionTime":"2025-12-10T09:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.957660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.957702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.957716 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.957734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:29 crc kubenswrapper[4715]: I1210 09:34:29.957756 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:29Z","lastTransitionTime":"2025-12-10T09:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.060122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.060158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.060166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.060187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.060203 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:30Z","lastTransitionTime":"2025-12-10T09:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.075250 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.090160 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/0.log" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.092266 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.092955 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa" exitCode=1 Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.093004 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.093687 4715 scope.go:117] "RemoveContainer" containerID="2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.107142 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.118202 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.131418 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.144456 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.157913 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.162129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.162159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.162169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.162189 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.162199 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:30Z","lastTransitionTime":"2025-12-10T09:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.174030 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.184231 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.196732 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.214468 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.224904 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: E1210 09:34:30.227320 4715 secret.go:188] Couldn't get secret openshift-ovn-kubernetes/ovn-control-plane-metrics-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 09:34:30 crc kubenswrapper[4715]: E1210 09:34:30.227406 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc3c82fe-6594-41d0-96d5-d8977ac76b15-ovn-control-plane-metrics-cert podName:cc3c82fe-6594-41d0-96d5-d8977ac76b15 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:30.72738322 +0000 UTC m=+33.470929481 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovn-control-plane-metrics-cert" (UniqueName: "kubernetes.io/secret/cc3c82fe-6594-41d0-96d5-d8977ac76b15-ovn-control-plane-metrics-cert") pod "ovnkube-control-plane-749d76644c-rnj27" (UID: "cc3c82fe-6594-41d0-96d5-d8977ac76b15") : failed to sync secret cache: timed out waiting for the condition Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.233837 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.245083 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.254628 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.264615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.264645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.264654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.264671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.264679 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:30Z","lastTransitionTime":"2025-12-10T09:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.265814 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.280763 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"ons/factory.go:141\\\\nI1210 09:34:28.554305 5959 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554328 5959 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554365 5959 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 09:34:28.554483 5959 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554736 5959 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 09:34:28.553686 5959 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1210 09:34:28.556454 5959 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1210 09:34:28.554762 5959 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554784 5959 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 09:34:28.554790 5959 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.295343 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.367174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.367207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.367218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.367234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.367245 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:30Z","lastTransitionTime":"2025-12-10T09:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.469116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.469167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.469178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.469193 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.469219 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:30Z","lastTransitionTime":"2025-12-10T09:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.571977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.572036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.572049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.572065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.572407 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:30Z","lastTransitionTime":"2025-12-10T09:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.582358 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-h5xdb"] Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.582894 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:30 crc kubenswrapper[4715]: E1210 09:34:30.582992 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.597255 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.610002 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.622972 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.635231 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.638036 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.638095 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5b5p\" (UniqueName: \"kubernetes.io/projected/f450d9cd-df1c-456e-861b-d91b2b683417-kube-api-access-b5b5p\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.646700 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.655881 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.667745 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.674668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.674709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.674721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.674737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.674767 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:30Z","lastTransitionTime":"2025-12-10T09:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.679783 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.684076 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.690054 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.705054 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.739451 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.739518 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc3c82fe-6594-41d0-96d5-d8977ac76b15-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.739594 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5b5p\" (UniqueName: \"kubernetes.io/projected/f450d9cd-df1c-456e-861b-d91b2b683417-kube-api-access-b5b5p\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:30 crc kubenswrapper[4715]: E1210 09:34:30.739620 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:30 crc kubenswrapper[4715]: E1210 09:34:30.739707 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs podName:f450d9cd-df1c-456e-861b-d91b2b683417 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:31.239684921 +0000 UTC m=+33.983231182 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs") pod "network-metrics-daemon-h5xdb" (UID: "f450d9cd-df1c-456e-861b-d91b2b683417") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.745263 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.748483 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cc3c82fe-6594-41d0-96d5-d8977ac76b15-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rnj27\" (UID: \"cc3c82fe-6594-41d0-96d5-d8977ac76b15\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.765610 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5b5p\" (UniqueName: \"kubernetes.io/projected/f450d9cd-df1c-456e-861b-d91b2b683417-kube-api-access-b5b5p\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.771516 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.777367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.777398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.777408 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.777425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.777437 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:30Z","lastTransitionTime":"2025-12-10T09:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.784781 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.802483 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"ons/factory.go:141\\\\nI1210 09:34:28.554305 5959 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554328 5959 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554365 5959 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 09:34:28.554483 5959 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554736 5959 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 09:34:28.553686 5959 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1210 09:34:28.556454 5959 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1210 09:34:28.554762 5959 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554784 5959 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 09:34:28.554790 5959 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.812947 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.824470 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.837677 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.879836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.879903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.879925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.879967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.879982 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:30Z","lastTransitionTime":"2025-12-10T09:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.913054 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" Dec 10 09:34:30 crc kubenswrapper[4715]: W1210 09:34:30.928068 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc3c82fe_6594_41d0_96d5_d8977ac76b15.slice/crio-359c2077e78536027232a80c2a67f1885d71b3872ebd0821737607156b49f7f9 WatchSource:0}: Error finding container 359c2077e78536027232a80c2a67f1885d71b3872ebd0821737607156b49f7f9: Status 404 returned error can't find the container with id 359c2077e78536027232a80c2a67f1885d71b3872ebd0821737607156b49f7f9 Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.981950 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.982012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.982021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.982035 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:30 crc kubenswrapper[4715]: I1210 09:34:30.982044 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:30Z","lastTransitionTime":"2025-12-10T09:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.086035 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.086078 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.086088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.086107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.086119 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:31Z","lastTransitionTime":"2025-12-10T09:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.101618 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/0.log" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.104284 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.105261 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.105443 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.108977 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" event={"ID":"cc3c82fe-6594-41d0-96d5-d8977ac76b15","Type":"ContainerStarted","Data":"359c2077e78536027232a80c2a67f1885d71b3872ebd0821737607156b49f7f9"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.125545 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.137977 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.150818 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.165813 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.177343 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.187838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.187871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.187881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.187896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.187904 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:31Z","lastTransitionTime":"2025-12-10T09:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.188425 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.199519 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.216153 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"ons/factory.go:141\\\\nI1210 09:34:28.554305 5959 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554328 5959 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554365 5959 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 09:34:28.554483 5959 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554736 5959 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 09:34:28.553686 5959 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1210 09:34:28.556454 5959 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1210 09:34:28.554762 5959 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554784 5959 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 09:34:28.554790 5959 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.225737 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.235448 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.245299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.245495 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.245573 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs podName:f450d9cd-df1c-456e-861b-d91b2b683417 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:32.245553435 +0000 UTC m=+34.989099686 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs") pod "network-metrics-daemon-h5xdb" (UID: "f450d9cd-df1c-456e-861b-d91b2b683417") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.247595 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.256302 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.277659 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.290790 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.290839 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.290851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.290870 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.290882 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:31Z","lastTransitionTime":"2025-12-10T09:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.295114 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.308865 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.320906 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.331493 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.393884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.393959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.393969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.393990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.394008 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:31Z","lastTransitionTime":"2025-12-10T09:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.446408 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.446556 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:34:47.446536169 +0000 UTC m=+50.190082420 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.498287 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.498365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.498393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.498424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.498450 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:31Z","lastTransitionTime":"2025-12-10T09:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.547238 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.547279 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.547297 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.547331 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547441 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547495 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547556 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:47.54752434 +0000 UTC m=+50.291070631 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547585 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:47.547571232 +0000 UTC m=+50.291117523 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547589 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547608 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547618 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547651 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:47.547633724 +0000 UTC m=+50.291179975 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547451 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547674 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547680 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.547698 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:47.547692275 +0000 UTC m=+50.291238526 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.602447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.602479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.602506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.602529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.602540 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:31Z","lastTransitionTime":"2025-12-10T09:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.604569 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.604637 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.604706 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.604778 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.604878 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:31 crc kubenswrapper[4715]: E1210 09:34:31.605044 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.705225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.705270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.705278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.705293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.705303 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:31Z","lastTransitionTime":"2025-12-10T09:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.808043 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.808086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.808097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.808113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.808126 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:31Z","lastTransitionTime":"2025-12-10T09:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.910577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.911030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.911048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.911073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:31 crc kubenswrapper[4715]: I1210 09:34:31.911090 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:31Z","lastTransitionTime":"2025-12-10T09:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.013876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.013957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.013969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.013990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.014001 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:32Z","lastTransitionTime":"2025-12-10T09:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.117690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.117728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.117741 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.117761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.117773 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:32Z","lastTransitionTime":"2025-12-10T09:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.119725 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" event={"ID":"cc3c82fe-6594-41d0-96d5-d8977ac76b15","Type":"ContainerStarted","Data":"ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.119794 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" event={"ID":"cc3c82fe-6594-41d0-96d5-d8977ac76b15","Type":"ContainerStarted","Data":"79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.123694 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/1.log" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.126108 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/0.log" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.130384 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.131356 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3" exitCode=1 Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.131405 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.131446 4715 scope.go:117] "RemoveContainer" containerID="2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.132307 4715 scope.go:117] "RemoveContainer" containerID="e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3" Dec 10 09:34:32 crc kubenswrapper[4715]: E1210 09:34:32.132488 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.156006 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"ons/factory.go:141\\\\nI1210 09:34:28.554305 5959 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554328 5959 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554365 5959 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 09:34:28.554483 5959 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554736 5959 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 09:34:28.553686 5959 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1210 09:34:28.556454 5959 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1210 09:34:28.554762 5959 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554784 5959 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 09:34:28.554790 5959 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.170023 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.183313 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.198465 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.214317 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.220127 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.220199 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.220215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.220239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.220254 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:32Z","lastTransitionTime":"2025-12-10T09:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.229407 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.243066 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.254354 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:32 crc kubenswrapper[4715]: E1210 09:34:32.254683 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:32 crc kubenswrapper[4715]: E1210 09:34:32.254832 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs podName:f450d9cd-df1c-456e-861b-d91b2b683417 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:34.25477686 +0000 UTC m=+36.998323121 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs") pod "network-metrics-daemon-h5xdb" (UID: "f450d9cd-df1c-456e-861b-d91b2b683417") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.255393 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.268624 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.279947 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.291546 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.306401 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.321342 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.323727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.323782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.323797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.323817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.323830 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:32Z","lastTransitionTime":"2025-12-10T09:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.338561 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.357891 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.369675 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.381235 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.397888 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.414384 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.425581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.425614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.425623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.425635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.425645 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:32Z","lastTransitionTime":"2025-12-10T09:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.435322 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c9ad654b7e2da2a49cb0c80c2debabd37d59b35bff4b6dfaab9b977ee8623fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"message\\\":\\\"ons/factory.go:141\\\\nI1210 09:34:28.554305 5959 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554328 5959 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554365 5959 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 09:34:28.554483 5959 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554736 5959 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 09:34:28.553686 5959 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1210 09:34:28.556454 5959 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1210 09:34:28.554762 5959 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 09:34:28.554784 5959 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 09:34:28.554790 5959 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:31Z\\\",\\\"message\\\":\\\"re timer (time: 2025-12-10 09:34:32.623439535 +0000 UTC m=+1.884246200): skip\\\\nI1210 09:34:31.505481 6161 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 73.092µs)\\\\nI1210 09:34:31.505502 6161 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 09:34:31.505521 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 09:34:31.505538 6161 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 09:34:31.505635 6161 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z]\\\\nI1210 09:34:31.505697 6161 event.go:377] Event(v1.ObjectRefere\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.446700 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.457387 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.469852 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.483860 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.497639 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.509429 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.521187 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.528530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.528569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.528578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.528594 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.528604 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:32Z","lastTransitionTime":"2025-12-10T09:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.535554 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.545606 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.555886 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.577038 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.588522 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.599250 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.604010 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:32 crc kubenswrapper[4715]: E1210 09:34:32.604131 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.612854 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:32Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.630425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.630471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.630486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.630504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.630514 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:32Z","lastTransitionTime":"2025-12-10T09:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.733079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.733157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.733168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.733184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.733194 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:32Z","lastTransitionTime":"2025-12-10T09:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.836462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.836520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.836530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.836551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.836562 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:32Z","lastTransitionTime":"2025-12-10T09:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.939327 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.939365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.939374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.939387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:32 crc kubenswrapper[4715]: I1210 09:34:32.939397 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:32Z","lastTransitionTime":"2025-12-10T09:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.043142 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.043223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.043249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.043277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.043298 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:33Z","lastTransitionTime":"2025-12-10T09:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.137848 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/1.log" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.142602 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.145708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.145760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.145773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.145798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.145810 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:33Z","lastTransitionTime":"2025-12-10T09:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.248777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.248814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.248822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.248838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.248848 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:33Z","lastTransitionTime":"2025-12-10T09:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.351948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.352048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.352065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.352085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.352098 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:33Z","lastTransitionTime":"2025-12-10T09:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.455494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.455550 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.455562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.455586 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.455602 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:33Z","lastTransitionTime":"2025-12-10T09:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.558668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.558742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.558752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.558772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.558782 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:33Z","lastTransitionTime":"2025-12-10T09:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.604694 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:33 crc kubenswrapper[4715]: E1210 09:34:33.604878 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.606173 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:33 crc kubenswrapper[4715]: E1210 09:34:33.606327 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.606392 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:33 crc kubenswrapper[4715]: E1210 09:34:33.606443 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.661984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.662049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.662066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.662090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.662107 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:33Z","lastTransitionTime":"2025-12-10T09:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.765595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.766048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.766245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.766418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.766638 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:33Z","lastTransitionTime":"2025-12-10T09:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.874645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.874724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.874747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.874778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.874806 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:33Z","lastTransitionTime":"2025-12-10T09:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.977414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.977511 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.977534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.977566 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:33 crc kubenswrapper[4715]: I1210 09:34:33.977590 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:33Z","lastTransitionTime":"2025-12-10T09:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.080669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.080752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.080787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.080807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.080819 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:34Z","lastTransitionTime":"2025-12-10T09:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.183155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.183200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.183216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.183236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.183251 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:34Z","lastTransitionTime":"2025-12-10T09:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.276078 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:34 crc kubenswrapper[4715]: E1210 09:34:34.276214 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:34 crc kubenswrapper[4715]: E1210 09:34:34.276266 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs podName:f450d9cd-df1c-456e-861b-d91b2b683417 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:38.276249498 +0000 UTC m=+41.019795749 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs") pod "network-metrics-daemon-h5xdb" (UID: "f450d9cd-df1c-456e-861b-d91b2b683417") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.284982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.285026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.285043 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.285065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.285081 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:34Z","lastTransitionTime":"2025-12-10T09:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.388328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.388376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.388404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.388417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.388427 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:34Z","lastTransitionTime":"2025-12-10T09:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.491274 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.491320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.491331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.491356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.491390 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:34Z","lastTransitionTime":"2025-12-10T09:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.593971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.594012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.594023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.594041 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.594052 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:34Z","lastTransitionTime":"2025-12-10T09:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.604047 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:34 crc kubenswrapper[4715]: E1210 09:34:34.604178 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.697357 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.697414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.697429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.697450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.697465 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:34Z","lastTransitionTime":"2025-12-10T09:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.800324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.800380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.800396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.800420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.800436 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:34Z","lastTransitionTime":"2025-12-10T09:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.903132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.903176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.903188 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.903209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:34 crc kubenswrapper[4715]: I1210 09:34:34.903220 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:34Z","lastTransitionTime":"2025-12-10T09:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.006631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.006689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.006703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.006726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.006750 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:35Z","lastTransitionTime":"2025-12-10T09:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.109490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.109524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.109537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.109550 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.109559 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:35Z","lastTransitionTime":"2025-12-10T09:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.211664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.211710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.211722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.211739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.211749 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:35Z","lastTransitionTime":"2025-12-10T09:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.314768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.314822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.314837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.314855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.314866 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:35Z","lastTransitionTime":"2025-12-10T09:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.417386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.417457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.417467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.417483 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.417495 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:35Z","lastTransitionTime":"2025-12-10T09:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.519881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.519955 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.519967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.519984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.519995 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:35Z","lastTransitionTime":"2025-12-10T09:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.604959 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:35 crc kubenswrapper[4715]: E1210 09:34:35.605121 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.604966 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:35 crc kubenswrapper[4715]: E1210 09:34:35.605209 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.605204 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:35 crc kubenswrapper[4715]: E1210 09:34:35.605270 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.622271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.622309 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.622320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.622335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.622345 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:35Z","lastTransitionTime":"2025-12-10T09:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.724720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.724772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.724786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.724807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.724823 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:35Z","lastTransitionTime":"2025-12-10T09:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.827844 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.827971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.827997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.828027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.828049 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:35Z","lastTransitionTime":"2025-12-10T09:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.931778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.931854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.931876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.931904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:35 crc kubenswrapper[4715]: I1210 09:34:35.931973 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:35Z","lastTransitionTime":"2025-12-10T09:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.034354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.034429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.034452 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.034482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.034500 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:36Z","lastTransitionTime":"2025-12-10T09:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.136555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.136603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.136613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.136626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.136635 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:36Z","lastTransitionTime":"2025-12-10T09:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.238257 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.238288 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.238299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.238315 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.238325 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:36Z","lastTransitionTime":"2025-12-10T09:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.341546 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.341596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.341610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.341629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.341642 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:36Z","lastTransitionTime":"2025-12-10T09:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.444999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.445061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.445079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.445104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.445121 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:36Z","lastTransitionTime":"2025-12-10T09:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.548765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.548813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.548824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.548844 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.548855 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:36Z","lastTransitionTime":"2025-12-10T09:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.604141 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:36 crc kubenswrapper[4715]: E1210 09:34:36.604315 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.651879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.651940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.651951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.651968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.651979 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:36Z","lastTransitionTime":"2025-12-10T09:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.754145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.754194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.754205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.754217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.754227 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:36Z","lastTransitionTime":"2025-12-10T09:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.855951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.855982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.855993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.856008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.856016 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:36Z","lastTransitionTime":"2025-12-10T09:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.959938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.959981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.959995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.960013 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:36 crc kubenswrapper[4715]: I1210 09:34:36.960029 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:36Z","lastTransitionTime":"2025-12-10T09:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.063551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.063626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.063650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.063678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.063701 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:37Z","lastTransitionTime":"2025-12-10T09:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.166426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.166473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.166482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.166495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.166504 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:37Z","lastTransitionTime":"2025-12-10T09:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.269162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.269200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.269216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.269233 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.269243 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:37Z","lastTransitionTime":"2025-12-10T09:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.352520 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.353712 4715 scope.go:117] "RemoveContainer" containerID="e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3" Dec 10 09:34:37 crc kubenswrapper[4715]: E1210 09:34:37.354069 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.372957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.374007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.374048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.374070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.374083 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:37Z","lastTransitionTime":"2025-12-10T09:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.374636 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.394569 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.414605 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.425587 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.439833 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.457632 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.471240 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.475979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.476045 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.476062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.476081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.476096 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:37Z","lastTransitionTime":"2025-12-10T09:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.485592 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.504509 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.526383 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.538374 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.548070 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.565133 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:31Z\\\",\\\"message\\\":\\\"re timer (time: 2025-12-10 09:34:32.623439535 +0000 UTC m=+1.884246200): skip\\\\nI1210 09:34:31.505481 6161 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 73.092µs)\\\\nI1210 09:34:31.505502 6161 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 09:34:31.505521 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 09:34:31.505538 6161 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 09:34:31.505635 6161 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z]\\\\nI1210 09:34:31.505697 6161 event.go:377] Event(v1.ObjectRefere\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.577936 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.579034 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.579063 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.579071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.579085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.579098 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:37Z","lastTransitionTime":"2025-12-10T09:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.589404 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.602786 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.604001 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.604120 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:37 crc kubenswrapper[4715]: E1210 09:34:37.604326 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.604419 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:37 crc kubenswrapper[4715]: E1210 09:34:37.604853 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:37 crc kubenswrapper[4715]: E1210 09:34:37.605019 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.614768 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.624751 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.636981 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.657006 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:31Z\\\",\\\"message\\\":\\\"re timer (time: 2025-12-10 09:34:32.623439535 +0000 UTC m=+1.884246200): skip\\\\nI1210 09:34:31.505481 6161 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 73.092µs)\\\\nI1210 09:34:31.505502 6161 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 09:34:31.505521 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 09:34:31.505538 6161 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 09:34:31.505635 6161 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z]\\\\nI1210 09:34:31.505697 6161 event.go:377] Event(v1.ObjectRefere\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.668103 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.676907 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.680837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.680886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.680901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.680946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.680963 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:37Z","lastTransitionTime":"2025-12-10T09:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.689517 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.702306 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.716825 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.730519 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.748794 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.764186 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.774993 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.783643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.783699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.783707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.783721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.783731 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:37Z","lastTransitionTime":"2025-12-10T09:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.786365 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.805174 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.818112 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.829376 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.842233 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.886513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.886551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.886561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.886577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.886587 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:37Z","lastTransitionTime":"2025-12-10T09:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.988948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.988984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.988994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.989009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:37 crc kubenswrapper[4715]: I1210 09:34:37.989019 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:37Z","lastTransitionTime":"2025-12-10T09:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.090726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.090779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.090797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.090819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.090836 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.193822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.193870 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.193879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.193894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.193905 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.296443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.296520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.296539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.296568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.296585 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.316746 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:38 crc kubenswrapper[4715]: E1210 09:34:38.316873 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:38 crc kubenswrapper[4715]: E1210 09:34:38.316946 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs podName:f450d9cd-df1c-456e-861b-d91b2b683417 nodeName:}" failed. No retries permitted until 2025-12-10 09:34:46.31692959 +0000 UTC m=+49.060475841 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs") pod "network-metrics-daemon-h5xdb" (UID: "f450d9cd-df1c-456e-861b-d91b2b683417") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.399238 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.399292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.399313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.399339 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.399358 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.502206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.502274 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.502298 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.502319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.502332 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.604351 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:38 crc kubenswrapper[4715]: E1210 09:34:38.604526 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.608524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.608566 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.608581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.608598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.608611 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.710936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.711193 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.711205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.711217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.711225 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.813385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.813429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.813438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.813453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.813463 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.916010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.916055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.916071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.916092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.916107 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.933260 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.933324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.933340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.933363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.933378 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: E1210 09:34:38.952260 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:38Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.957573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.957627 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.957639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.957664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.957677 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:38 crc kubenswrapper[4715]: E1210 09:34:38.977033 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:38Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.986941 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.987052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.988109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.988259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:38 crc kubenswrapper[4715]: I1210 09:34:38.988286 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:38Z","lastTransitionTime":"2025-12-10T09:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: E1210 09:34:39.004250 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:39Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.009274 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.009314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.009324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.009340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.009349 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: E1210 09:34:39.024668 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:39Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.029768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.029823 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.029836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.029855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.029867 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: E1210 09:34:39.043081 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:39Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:39 crc kubenswrapper[4715]: E1210 09:34:39.043352 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.045249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.045330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.045347 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.045371 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.045389 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.149301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.149385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.149411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.149442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.149461 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.252062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.252113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.252124 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.252140 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.252150 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.354865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.354935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.354946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.354963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.354975 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.457521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.457556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.457567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.457585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.457600 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.559536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.559602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.559616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.559638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.559654 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.604256 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.604345 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.604457 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:39 crc kubenswrapper[4715]: E1210 09:34:39.604443 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:39 crc kubenswrapper[4715]: E1210 09:34:39.604624 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:39 crc kubenswrapper[4715]: E1210 09:34:39.604716 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.662653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.662688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.662701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.662717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.662729 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.765254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.765474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.765557 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.765648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.765766 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.869185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.869239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.869249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.869268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.869281 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.971848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.971896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.971906 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.971945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:39 crc kubenswrapper[4715]: I1210 09:34:39.971963 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:39Z","lastTransitionTime":"2025-12-10T09:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.074577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.074628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.074641 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.074661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.074676 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:40Z","lastTransitionTime":"2025-12-10T09:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.177010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.177079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.177107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.177127 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.177144 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:40Z","lastTransitionTime":"2025-12-10T09:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.280217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.280291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.280318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.280347 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.280368 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:40Z","lastTransitionTime":"2025-12-10T09:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.383391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.383432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.383442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.383457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.383467 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:40Z","lastTransitionTime":"2025-12-10T09:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.492428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.492468 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.492478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.492496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.492507 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:40Z","lastTransitionTime":"2025-12-10T09:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.595024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.595051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.595061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.595075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.595085 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:40Z","lastTransitionTime":"2025-12-10T09:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.604529 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:40 crc kubenswrapper[4715]: E1210 09:34:40.604662 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.697864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.697925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.697944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.697963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.697975 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:40Z","lastTransitionTime":"2025-12-10T09:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.799959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.800255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.800348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.800487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.800570 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:40Z","lastTransitionTime":"2025-12-10T09:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.903716 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.903765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.903774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.903787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:40 crc kubenswrapper[4715]: I1210 09:34:40.903797 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:40Z","lastTransitionTime":"2025-12-10T09:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.007223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.007308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.007335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.007370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.007393 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:41Z","lastTransitionTime":"2025-12-10T09:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.110182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.110249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.110285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.110317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.110342 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:41Z","lastTransitionTime":"2025-12-10T09:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.212978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.213051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.213077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.213121 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.213143 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:41Z","lastTransitionTime":"2025-12-10T09:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.316494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.316542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.316621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.316647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.316665 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:41Z","lastTransitionTime":"2025-12-10T09:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.419341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.419393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.419405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.419423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.419434 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:41Z","lastTransitionTime":"2025-12-10T09:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.522083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.522143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.522159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.522180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.522195 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:41Z","lastTransitionTime":"2025-12-10T09:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.604112 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.604154 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:41 crc kubenswrapper[4715]: E1210 09:34:41.604298 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.604318 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:41 crc kubenswrapper[4715]: E1210 09:34:41.604456 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:41 crc kubenswrapper[4715]: E1210 09:34:41.604536 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.625259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.625540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.625627 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.625762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.625864 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:41Z","lastTransitionTime":"2025-12-10T09:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.728136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.728166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.728174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.728188 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.728198 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:41Z","lastTransitionTime":"2025-12-10T09:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.831076 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.831136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.831149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.831170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.831189 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:41Z","lastTransitionTime":"2025-12-10T09:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.935052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.935115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.935130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.935157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:41 crc kubenswrapper[4715]: I1210 09:34:41.935174 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:41Z","lastTransitionTime":"2025-12-10T09:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.038198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.038269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.038293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.038323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.038345 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:42Z","lastTransitionTime":"2025-12-10T09:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.141822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.141944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.141969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.141998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.142015 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:42Z","lastTransitionTime":"2025-12-10T09:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.245488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.245609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.245628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.245659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.245677 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:42Z","lastTransitionTime":"2025-12-10T09:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.348792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.348847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.348864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.348887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.348905 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:42Z","lastTransitionTime":"2025-12-10T09:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.451383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.451428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.451441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.451457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.451469 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:42Z","lastTransitionTime":"2025-12-10T09:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.555113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.555209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.555245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.555279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.555301 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:42Z","lastTransitionTime":"2025-12-10T09:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.604467 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:42 crc kubenswrapper[4715]: E1210 09:34:42.604654 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.658347 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.658405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.658417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.658440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.658454 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:42Z","lastTransitionTime":"2025-12-10T09:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.760808 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.760848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.760859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.760876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.760886 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:42Z","lastTransitionTime":"2025-12-10T09:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.863967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.864055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.864077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.864105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.864130 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:42Z","lastTransitionTime":"2025-12-10T09:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.966843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.966907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.966982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.967013 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:42 crc kubenswrapper[4715]: I1210 09:34:42.967036 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:42Z","lastTransitionTime":"2025-12-10T09:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.069665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.069733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.069753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.069782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.069803 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:43Z","lastTransitionTime":"2025-12-10T09:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.172710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.172793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.172819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.172848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.172869 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:43Z","lastTransitionTime":"2025-12-10T09:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.276580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.276681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.276708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.276745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.276785 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:43Z","lastTransitionTime":"2025-12-10T09:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.380387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.380452 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.380471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.380495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.380516 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:43Z","lastTransitionTime":"2025-12-10T09:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.484396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.484478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.484498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.484831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.484855 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:43Z","lastTransitionTime":"2025-12-10T09:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.588282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.588772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.588963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.589145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.589298 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:43Z","lastTransitionTime":"2025-12-10T09:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.604687 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:43 crc kubenswrapper[4715]: E1210 09:34:43.604809 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.604893 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.604893 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:43 crc kubenswrapper[4715]: E1210 09:34:43.605184 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:43 crc kubenswrapper[4715]: E1210 09:34:43.605349 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.691453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.691485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.691492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.691506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.691514 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:43Z","lastTransitionTime":"2025-12-10T09:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.794090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.794169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.794180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.794194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.794204 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:43Z","lastTransitionTime":"2025-12-10T09:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.897709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.898147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.898328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.898580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:43 crc kubenswrapper[4715]: I1210 09:34:43.898767 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:43Z","lastTransitionTime":"2025-12-10T09:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.002529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.002590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.002603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.002625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.002638 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:44Z","lastTransitionTime":"2025-12-10T09:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.105348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.105401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.105412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.105431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.105443 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:44Z","lastTransitionTime":"2025-12-10T09:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.208495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.208559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.208574 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.208593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.208605 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:44Z","lastTransitionTime":"2025-12-10T09:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.311626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.311665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.311675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.311689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.311699 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:44Z","lastTransitionTime":"2025-12-10T09:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.415046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.415090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.415101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.415115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.415125 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:44Z","lastTransitionTime":"2025-12-10T09:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.518394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.518485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.518506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.518529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.518550 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:44Z","lastTransitionTime":"2025-12-10T09:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.604372 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:44 crc kubenswrapper[4715]: E1210 09:34:44.604527 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.620687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.620740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.620756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.620781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.620805 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:44Z","lastTransitionTime":"2025-12-10T09:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.723661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.723695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.723705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.723719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.723729 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:44Z","lastTransitionTime":"2025-12-10T09:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.825840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.825884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.825900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.825942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.825964 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:44Z","lastTransitionTime":"2025-12-10T09:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.929617 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.929664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.929677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.929694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:44 crc kubenswrapper[4715]: I1210 09:34:44.929705 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:44Z","lastTransitionTime":"2025-12-10T09:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.033018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.033075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.033090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.033112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.033128 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:45Z","lastTransitionTime":"2025-12-10T09:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.135676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.135747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.135771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.135805 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.135827 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:45Z","lastTransitionTime":"2025-12-10T09:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.238706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.238754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.238765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.238780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.238790 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:45Z","lastTransitionTime":"2025-12-10T09:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.342615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.342670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.342687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.342708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.342725 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:45Z","lastTransitionTime":"2025-12-10T09:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.445630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.445678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.445695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.445717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.445734 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:45Z","lastTransitionTime":"2025-12-10T09:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.548949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.549006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.549023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.549045 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.549062 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:45Z","lastTransitionTime":"2025-12-10T09:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.604148 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.604234 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:45 crc kubenswrapper[4715]: E1210 09:34:45.604335 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:45 crc kubenswrapper[4715]: E1210 09:34:45.604432 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.604543 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:45 crc kubenswrapper[4715]: E1210 09:34:45.604717 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.651723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.651791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.651814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.651842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.651861 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:45Z","lastTransitionTime":"2025-12-10T09:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.754754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.754797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.754805 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.754819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.754828 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:45Z","lastTransitionTime":"2025-12-10T09:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.857427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.857496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.857514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.857538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.857555 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:45Z","lastTransitionTime":"2025-12-10T09:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.960842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.960963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.960984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.961010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:45 crc kubenswrapper[4715]: I1210 09:34:45.961030 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:45Z","lastTransitionTime":"2025-12-10T09:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.064474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.064540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.064552 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.064577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.064594 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:46Z","lastTransitionTime":"2025-12-10T09:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.169027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.169057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.169065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.169096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.169107 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:46Z","lastTransitionTime":"2025-12-10T09:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.271401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.271481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.271506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.271538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.271561 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:46Z","lastTransitionTime":"2025-12-10T09:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.374379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.374443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.374461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.374488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.374507 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:46Z","lastTransitionTime":"2025-12-10T09:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.405811 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:46 crc kubenswrapper[4715]: E1210 09:34:46.406037 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:46 crc kubenswrapper[4715]: E1210 09:34:46.406165 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs podName:f450d9cd-df1c-456e-861b-d91b2b683417 nodeName:}" failed. No retries permitted until 2025-12-10 09:35:02.40613152 +0000 UTC m=+65.149677811 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs") pod "network-metrics-daemon-h5xdb" (UID: "f450d9cd-df1c-456e-861b-d91b2b683417") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.477365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.477418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.477430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.477451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.477470 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:46Z","lastTransitionTime":"2025-12-10T09:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.580644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.580684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.580693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.580709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.580720 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:46Z","lastTransitionTime":"2025-12-10T09:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.604290 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:46 crc kubenswrapper[4715]: E1210 09:34:46.604454 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.683033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.683060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.683070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.683085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.683095 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:46Z","lastTransitionTime":"2025-12-10T09:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.784990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.785267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.785372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.785448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.785528 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:46Z","lastTransitionTime":"2025-12-10T09:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.888135 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.888379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.888466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.888585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.888669 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:46Z","lastTransitionTime":"2025-12-10T09:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.990610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.990813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.990900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.990984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:46 crc kubenswrapper[4715]: I1210 09:34:46.991101 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:46Z","lastTransitionTime":"2025-12-10T09:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.094375 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.094426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.094441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.094457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.094470 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:47Z","lastTransitionTime":"2025-12-10T09:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.197901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.198330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.198465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.198590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.198722 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:47Z","lastTransitionTime":"2025-12-10T09:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.302476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.302526 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.302542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.302565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.302586 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:47Z","lastTransitionTime":"2025-12-10T09:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.405305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.405847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.405952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.406017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.406074 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:47Z","lastTransitionTime":"2025-12-10T09:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.508323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.508550 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.508635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.508751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.508867 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:47Z","lastTransitionTime":"2025-12-10T09:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.519660 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.519804 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:35:19.519787476 +0000 UTC m=+82.263333727 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.605246 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.605444 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.605827 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.605964 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.606146 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.606245 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.614881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.614998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.615028 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.615060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.615086 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:47Z","lastTransitionTime":"2025-12-10T09:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.620252 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.620292 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.620311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.620351 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620426 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620481 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620497 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620493 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620532 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620546 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620575 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620507 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620512 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:35:19.620484918 +0000 UTC m=+82.364031219 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620685 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:35:19.620669274 +0000 UTC m=+82.364215525 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620698 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 09:35:19.620691784 +0000 UTC m=+82.364238035 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:47 crc kubenswrapper[4715]: E1210 09:34:47.620710 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 09:35:19.620704245 +0000 UTC m=+82.364250496 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.623829 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.638387 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.654504 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.667987 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.700311 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.722651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.722707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.722719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.722737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.722749 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:47Z","lastTransitionTime":"2025-12-10T09:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.731874 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.749574 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.763006 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.779756 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.790655 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.799584 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.814648 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.824952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.824989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.824999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.825016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.825026 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:47Z","lastTransitionTime":"2025-12-10T09:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.825224 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.836627 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.849355 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.861467 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.879047 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:31Z\\\",\\\"message\\\":\\\"re timer (time: 2025-12-10 09:34:32.623439535 +0000 UTC m=+1.884246200): skip\\\\nI1210 09:34:31.505481 6161 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 73.092µs)\\\\nI1210 09:34:31.505502 6161 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 09:34:31.505521 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 09:34:31.505538 6161 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 09:34:31.505635 6161 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z]\\\\nI1210 09:34:31.505697 6161 event.go:377] Event(v1.ObjectRefere\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:47Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.927024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.927055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.927062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.927075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:47 crc kubenswrapper[4715]: I1210 09:34:47.927084 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:47Z","lastTransitionTime":"2025-12-10T09:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.030859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.030960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.030982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.031012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.031033 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:48Z","lastTransitionTime":"2025-12-10T09:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.134818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.134887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.134950 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.134995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.135012 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:48Z","lastTransitionTime":"2025-12-10T09:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.238677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.238782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.238811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.238856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.239020 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:48Z","lastTransitionTime":"2025-12-10T09:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.341418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.341459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.341474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.341492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.341504 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:48Z","lastTransitionTime":"2025-12-10T09:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.444134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.444259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.444281 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.444684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.445007 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:48Z","lastTransitionTime":"2025-12-10T09:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.547763 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.547838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.547862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.547896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.547967 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:48Z","lastTransitionTime":"2025-12-10T09:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.604125 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:48 crc kubenswrapper[4715]: E1210 09:34:48.604613 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.605231 4715 scope.go:117] "RemoveContainer" containerID="e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.652826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.652971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.652995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.653051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.653073 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:48Z","lastTransitionTime":"2025-12-10T09:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.755607 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.755673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.755687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.755706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.755740 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:48Z","lastTransitionTime":"2025-12-10T09:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.858541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.858581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.858589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.858604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.858614 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:48Z","lastTransitionTime":"2025-12-10T09:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.970213 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.970275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.970287 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.970308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:48 crc kubenswrapper[4715]: I1210 09:34:48.970322 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:48Z","lastTransitionTime":"2025-12-10T09:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.072736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.072777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.072785 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.072800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.072812 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.176047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.176090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.176099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.176151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.176202 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.200483 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/1.log" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.202579 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.203467 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.203908 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.217002 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.228232 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.247105 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:31Z\\\",\\\"message\\\":\\\"re timer (time: 2025-12-10 09:34:32.623439535 +0000 UTC m=+1.884246200): skip\\\\nI1210 09:34:31.505481 6161 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 73.092µs)\\\\nI1210 09:34:31.505502 6161 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 09:34:31.505521 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 09:34:31.505538 6161 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 09:34:31.505635 6161 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z]\\\\nI1210 09:34:31.505697 6161 event.go:377] Event(v1.ObjectRefere\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.259461 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.272491 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.278526 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.278565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.278576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.278594 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.278605 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.285683 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.297355 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.313137 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.327559 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.341194 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.356128 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.376517 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.389834 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.403178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.403213 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.403221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.403236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.403246 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.414251 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.420004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.420027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.420035 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.420049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.420057 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.426527 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: E1210 09:34:49.432118 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.435537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.435559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.435567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.435580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.435589 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.437610 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: E1210 09:34:49.446817 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.450167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.450201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.450210 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.450223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.450234 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.454168 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: E1210 09:34:49.460500 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.470149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.470179 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.470187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.470203 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.470213 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: E1210 09:34:49.480826 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.484350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.484406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.484416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.484433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.484447 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: E1210 09:34:49.494504 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:49Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:49 crc kubenswrapper[4715]: E1210 09:34:49.494624 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.505292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.505340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.505350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.505366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.505376 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.604839 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.604880 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:49 crc kubenswrapper[4715]: E1210 09:34:49.605017 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.604849 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:49 crc kubenswrapper[4715]: E1210 09:34:49.605156 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:49 crc kubenswrapper[4715]: E1210 09:34:49.605268 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.613640 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.613677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.613688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.613706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.613715 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.715435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.715463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.715472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.715484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.715492 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.819118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.819186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.819209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.819239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.819261 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.921542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.921579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.921589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.921604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:49 crc kubenswrapper[4715]: I1210 09:34:49.921615 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:49Z","lastTransitionTime":"2025-12-10T09:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.025013 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.025367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.025547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.025732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.025959 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:50Z","lastTransitionTime":"2025-12-10T09:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.129707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.129766 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.129788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.129816 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.129839 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:50Z","lastTransitionTime":"2025-12-10T09:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.210372 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/2.log" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.211350 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/1.log" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.215352 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.216835 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e" exitCode=1 Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.216900 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.217139 4715 scope.go:117] "RemoveContainer" containerID="e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.218185 4715 scope.go:117] "RemoveContainer" containerID="1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e" Dec 10 09:34:50 crc kubenswrapper[4715]: E1210 09:34:50.218519 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.233215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.233272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.233289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.233313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.233331 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:50Z","lastTransitionTime":"2025-12-10T09:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.237639 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.254757 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.277262 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.293548 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.305151 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.323371 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.335977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.336017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.336028 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.336044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.336056 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:50Z","lastTransitionTime":"2025-12-10T09:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.339443 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.358353 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.377049 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.395287 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.413844 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.427199 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.438798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.438869 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.438886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.438934 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.438953 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:50Z","lastTransitionTime":"2025-12-10T09:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.446759 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:31Z\\\",\\\"message\\\":\\\"re timer (time: 2025-12-10 09:34:32.623439535 +0000 UTC m=+1.884246200): skip\\\\nI1210 09:34:31.505481 6161 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 73.092µs)\\\\nI1210 09:34:31.505502 6161 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 09:34:31.505521 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 09:34:31.505538 6161 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 09:34:31.505635 6161 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z]\\\\nI1210 09:34:31.505697 6161 event.go:377] Event(v1.ObjectRefere\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.461027 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.471281 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.485083 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.499530 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.542000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.542055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.542071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.542094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.542110 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:50Z","lastTransitionTime":"2025-12-10T09:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.604986 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:50 crc kubenswrapper[4715]: E1210 09:34:50.605185 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.644718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.644765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.644776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.644792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.644804 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:50Z","lastTransitionTime":"2025-12-10T09:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.747876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.748130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.748225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.748325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.748400 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:50Z","lastTransitionTime":"2025-12-10T09:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.834497 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.847556 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.851622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.851682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.851707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.851737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.851759 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:50Z","lastTransitionTime":"2025-12-10T09:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.862086 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.877354 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.893733 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.922161 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.936698 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.952810 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.955436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.955484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.955498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.955520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.955537 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:50Z","lastTransitionTime":"2025-12-10T09:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.969972 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:50 crc kubenswrapper[4715]: I1210 09:34:50.994284 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:50Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.011368 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.024541 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.045392 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e58428d5f238de640b0ee1fd735289ec337f5a8635dd5836caa3c950311d3ea3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:31Z\\\",\\\"message\\\":\\\"re timer (time: 2025-12-10 09:34:32.623439535 +0000 UTC m=+1.884246200): skip\\\\nI1210 09:34:31.505481 6161 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 73.092µs)\\\\nI1210 09:34:31.505502 6161 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 09:34:31.505521 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 09:34:31.505538 6161 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 09:34:31.505635 6161 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:31Z is after 2025-08-24T17:21:41Z]\\\\nI1210 09:34:31.505697 6161 event.go:377] Event(v1.ObjectRefere\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.057430 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.058624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.058669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.058681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.058698 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.058710 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:51Z","lastTransitionTime":"2025-12-10T09:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.070326 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.082478 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.094951 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.105248 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.118725 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.161517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.161576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.161587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.161601 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.161610 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:51Z","lastTransitionTime":"2025-12-10T09:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.221653 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/2.log" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.224138 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.225709 4715 scope.go:117] "RemoveContainer" containerID="1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e" Dec 10 09:34:51 crc kubenswrapper[4715]: E1210 09:34:51.225897 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.238156 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.251252 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.264963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.265034 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.265048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.265066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.265078 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:51Z","lastTransitionTime":"2025-12-10T09:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.267962 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.281415 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.302520 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.314428 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.325651 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.338838 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.350100 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.359426 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.367185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.367215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.367223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.367235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.367244 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:51Z","lastTransitionTime":"2025-12-10T09:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.375161 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.389376 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.401770 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.413289 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.430066 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.441806 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.452929 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.465656 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:51Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.468990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.469018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.469025 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.469038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.469046 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:51Z","lastTransitionTime":"2025-12-10T09:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.572045 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.572118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.572130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.572183 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.572194 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:51Z","lastTransitionTime":"2025-12-10T09:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.604877 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.604940 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:51 crc kubenswrapper[4715]: E1210 09:34:51.605112 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.605167 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:51 crc kubenswrapper[4715]: E1210 09:34:51.605374 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:51 crc kubenswrapper[4715]: E1210 09:34:51.605493 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.675002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.675047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.675061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.675080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.675095 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:51Z","lastTransitionTime":"2025-12-10T09:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.777394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.777456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.777472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.777497 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.777514 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:51Z","lastTransitionTime":"2025-12-10T09:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.880517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.880630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.880657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.880686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.880708 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:51Z","lastTransitionTime":"2025-12-10T09:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.983982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.984254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.984334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.984409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:51 crc kubenswrapper[4715]: I1210 09:34:51.984477 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:51Z","lastTransitionTime":"2025-12-10T09:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.087415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.087952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.088375 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.088819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.089235 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:52Z","lastTransitionTime":"2025-12-10T09:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.191601 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.191626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.191634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.191646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.191655 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:52Z","lastTransitionTime":"2025-12-10T09:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.294261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.294301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.294311 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.294327 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.294337 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:52Z","lastTransitionTime":"2025-12-10T09:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.396565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.396593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.396600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.396613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.396621 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:52Z","lastTransitionTime":"2025-12-10T09:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.499878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.500584 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.500832 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.501112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.501353 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:52Z","lastTransitionTime":"2025-12-10T09:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.604383 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:52 crc kubenswrapper[4715]: E1210 09:34:52.604567 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.605326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.605553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.605818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.606053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.606262 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:52Z","lastTransitionTime":"2025-12-10T09:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.710422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.711010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.711089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.711154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.711227 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:52Z","lastTransitionTime":"2025-12-10T09:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.814296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.814338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.814350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.814368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.814379 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:52Z","lastTransitionTime":"2025-12-10T09:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.917410 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.917473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.917485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.917502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:52 crc kubenswrapper[4715]: I1210 09:34:52.917512 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:52Z","lastTransitionTime":"2025-12-10T09:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.019906 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.019980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.019991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.020008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.020022 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:53Z","lastTransitionTime":"2025-12-10T09:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.122405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.122457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.122471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.122490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.122501 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:53Z","lastTransitionTime":"2025-12-10T09:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.225613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.225676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.225685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.225701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.225711 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:53Z","lastTransitionTime":"2025-12-10T09:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.329437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.329478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.329487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.329501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.329510 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:53Z","lastTransitionTime":"2025-12-10T09:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.432644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.432703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.432713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.432736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.432750 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:53Z","lastTransitionTime":"2025-12-10T09:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.535949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.536002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.536015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.536031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.536043 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:53Z","lastTransitionTime":"2025-12-10T09:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.604214 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.604302 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.604224 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:53 crc kubenswrapper[4715]: E1210 09:34:53.604466 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:53 crc kubenswrapper[4715]: E1210 09:34:53.604705 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:53 crc kubenswrapper[4715]: E1210 09:34:53.604856 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.638664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.638718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.638734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.638753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.638766 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:53Z","lastTransitionTime":"2025-12-10T09:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.742416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.742487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.742505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.742531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.742550 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:53Z","lastTransitionTime":"2025-12-10T09:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.845402 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.845463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.845481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.845507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.845524 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:53Z","lastTransitionTime":"2025-12-10T09:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.948440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.948493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.948515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.948551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:53 crc kubenswrapper[4715]: I1210 09:34:53.948574 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:53Z","lastTransitionTime":"2025-12-10T09:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.052024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.052074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.052085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.052101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.052113 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:54Z","lastTransitionTime":"2025-12-10T09:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.155214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.155490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.155652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.155787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.155970 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:54Z","lastTransitionTime":"2025-12-10T09:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.258515 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.258567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.258578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.258598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.258612 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:54Z","lastTransitionTime":"2025-12-10T09:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.361361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.361404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.361412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.361427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.361435 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:54Z","lastTransitionTime":"2025-12-10T09:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.464031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.464077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.464089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.464106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.464117 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:54Z","lastTransitionTime":"2025-12-10T09:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.566750 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.566802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.566815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.566832 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.566841 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:54Z","lastTransitionTime":"2025-12-10T09:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.604571 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:54 crc kubenswrapper[4715]: E1210 09:34:54.605207 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.669681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.669745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.669768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.669798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.669824 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:54Z","lastTransitionTime":"2025-12-10T09:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.771851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.772224 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.772389 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.772537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.772654 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:54Z","lastTransitionTime":"2025-12-10T09:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.875344 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.875395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.875413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.875437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.875453 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:54Z","lastTransitionTime":"2025-12-10T09:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.978234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.978318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.978340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.978374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:54 crc kubenswrapper[4715]: I1210 09:34:54.978397 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:54Z","lastTransitionTime":"2025-12-10T09:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.080770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.081033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.081107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.081200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.081264 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:55Z","lastTransitionTime":"2025-12-10T09:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.183434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.183645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.183707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.183773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.183856 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:55Z","lastTransitionTime":"2025-12-10T09:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.286177 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.286225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.286235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.286248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.286257 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:55Z","lastTransitionTime":"2025-12-10T09:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.389163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.389730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.389829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.389945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.390058 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:55Z","lastTransitionTime":"2025-12-10T09:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.492540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.492587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.492602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.492621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.492634 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:55Z","lastTransitionTime":"2025-12-10T09:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.595733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.595800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.595811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.595828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.595838 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:55Z","lastTransitionTime":"2025-12-10T09:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.604298 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.604408 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:55 crc kubenswrapper[4715]: E1210 09:34:55.604435 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.604511 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:55 crc kubenswrapper[4715]: E1210 09:34:55.604593 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:55 crc kubenswrapper[4715]: E1210 09:34:55.604681 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.698287 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.698345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.698357 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.698380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.698393 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:55Z","lastTransitionTime":"2025-12-10T09:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.802057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.802126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.802137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.802158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.802175 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:55Z","lastTransitionTime":"2025-12-10T09:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.905561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.905653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.905676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.905715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:55 crc kubenswrapper[4715]: I1210 09:34:55.905741 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:55Z","lastTransitionTime":"2025-12-10T09:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.008537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.008582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.008594 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.008671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.008682 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:56Z","lastTransitionTime":"2025-12-10T09:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.111760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.111827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.111844 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.111869 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.111885 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:56Z","lastTransitionTime":"2025-12-10T09:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.214075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.214139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.214151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.214177 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.214190 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:56Z","lastTransitionTime":"2025-12-10T09:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.317812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.317865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.317875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.317896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.317928 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:56Z","lastTransitionTime":"2025-12-10T09:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.420881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.420943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.420954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.420973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.420983 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:56Z","lastTransitionTime":"2025-12-10T09:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.523444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.523493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.523505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.523520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.523532 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:56Z","lastTransitionTime":"2025-12-10T09:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.605195 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:56 crc kubenswrapper[4715]: E1210 09:34:56.605351 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.635450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.635500 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.635510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.635528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.635540 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:56Z","lastTransitionTime":"2025-12-10T09:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.738478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.738841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.738992 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.739089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.739173 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:56Z","lastTransitionTime":"2025-12-10T09:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.843981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.844386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.844459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.844527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.844584 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:56Z","lastTransitionTime":"2025-12-10T09:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.948116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.948175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.948193 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.948220 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:56 crc kubenswrapper[4715]: I1210 09:34:56.948240 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:56Z","lastTransitionTime":"2025-12-10T09:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.051351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.051737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.051857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.052020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.052159 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:57Z","lastTransitionTime":"2025-12-10T09:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.156473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.156524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.156538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.156560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.156578 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:57Z","lastTransitionTime":"2025-12-10T09:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.259093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.259524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.259640 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.260027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.260145 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:57Z","lastTransitionTime":"2025-12-10T09:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.363522 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.363606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.363620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.363652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.363679 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:57Z","lastTransitionTime":"2025-12-10T09:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.467308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.467380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.467396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.467421 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.467439 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:57Z","lastTransitionTime":"2025-12-10T09:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.570484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.570544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.570553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.570573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.570585 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:57Z","lastTransitionTime":"2025-12-10T09:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.604428 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.604483 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:57 crc kubenswrapper[4715]: E1210 09:34:57.604738 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.604827 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:57 crc kubenswrapper[4715]: E1210 09:34:57.605005 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:57 crc kubenswrapper[4715]: E1210 09:34:57.605827 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.622135 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.641891 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.660434 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.673794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.673835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.673848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.673866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.673879 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:57Z","lastTransitionTime":"2025-12-10T09:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.677666 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.689548 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.703549 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.716286 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.728241 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.743600 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.762296 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.775809 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.777379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.777404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.777413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.777429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.777440 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:57Z","lastTransitionTime":"2025-12-10T09:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.789959 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.810967 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.823749 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.836904 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.850863 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.867073 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.880299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.880350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.880362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.880380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.880393 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:57Z","lastTransitionTime":"2025-12-10T09:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.882294 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:57Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.983780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.983821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.983831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.983853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:57 crc kubenswrapper[4715]: I1210 09:34:57.983867 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:57Z","lastTransitionTime":"2025-12-10T09:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.087368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.087440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.087451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.087472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.087484 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:58Z","lastTransitionTime":"2025-12-10T09:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.189909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.190516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.190829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.191094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.191308 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:58Z","lastTransitionTime":"2025-12-10T09:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.294402 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.294464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.294481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.294508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.294527 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:58Z","lastTransitionTime":"2025-12-10T09:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.397755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.397811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.397828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.397851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.397868 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:58Z","lastTransitionTime":"2025-12-10T09:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.500449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.500706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.500770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.500837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.500905 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:58Z","lastTransitionTime":"2025-12-10T09:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.604379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.604435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.604493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.604522 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.604543 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:58Z","lastTransitionTime":"2025-12-10T09:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.608076 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:34:58 crc kubenswrapper[4715]: E1210 09:34:58.608353 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.711558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.711865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.711993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.712127 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.712238 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:58Z","lastTransitionTime":"2025-12-10T09:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.814980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.815428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.815517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.815593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.815660 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:58Z","lastTransitionTime":"2025-12-10T09:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.919096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.919142 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.919152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.919168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:58 crc kubenswrapper[4715]: I1210 09:34:58.919182 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:58Z","lastTransitionTime":"2025-12-10T09:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.021663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.022167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.022259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.022351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.022420 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.126473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.126836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.126972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.127089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.127181 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.230845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.230897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.230943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.230968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.230985 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.333808 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.333862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.333877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.333896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.333935 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.436514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.436546 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.436556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.436569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.436580 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.539441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.539470 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.539477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.539490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.539499 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.584428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.584483 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.584504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.584532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.584553 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.607676 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:34:59 crc kubenswrapper[4715]: E1210 09:34:59.607783 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.607966 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:34:59 crc kubenswrapper[4715]: E1210 09:34:59.608014 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.608111 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:34:59 crc kubenswrapper[4715]: E1210 09:34:59.608155 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:34:59 crc kubenswrapper[4715]: E1210 09:34:59.608523 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:59Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.611886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.611945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.611956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.611972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.611983 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: E1210 09:34:59.623811 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:59Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.626849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.626874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.626884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.626901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.626926 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: E1210 09:34:59.641425 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:59Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.645498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.645556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.645574 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.645600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.645619 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: E1210 09:34:59.660855 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:59Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.664435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.664471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.664480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.664495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.664505 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: E1210 09:34:59.677345 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:34:59Z is after 2025-08-24T17:21:41Z" Dec 10 09:34:59 crc kubenswrapper[4715]: E1210 09:34:59.677464 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.679010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.679072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.679084 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.679107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.679134 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.782235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.782289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.782300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.782319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.782331 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.884616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.884657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.884667 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.884682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.884693 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.987038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.987088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.987099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.987116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:34:59 crc kubenswrapper[4715]: I1210 09:34:59.987132 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:34:59Z","lastTransitionTime":"2025-12-10T09:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.089677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.089746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.089765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.089789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.089805 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:00Z","lastTransitionTime":"2025-12-10T09:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.193523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.193580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.193603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.193631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.193658 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:00Z","lastTransitionTime":"2025-12-10T09:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.298831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.298868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.298876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.298889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.298899 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:00Z","lastTransitionTime":"2025-12-10T09:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.401878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.402265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.402535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.402690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.402872 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:00Z","lastTransitionTime":"2025-12-10T09:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.507442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.507493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.507503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.507525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.507538 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:00Z","lastTransitionTime":"2025-12-10T09:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.604843 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:00 crc kubenswrapper[4715]: E1210 09:35:00.605046 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.609976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.610059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.610091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.610127 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.610156 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:00Z","lastTransitionTime":"2025-12-10T09:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.713736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.713777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.713787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.713804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.713818 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:00Z","lastTransitionTime":"2025-12-10T09:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.816427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.816474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.816484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.816502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.816514 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:00Z","lastTransitionTime":"2025-12-10T09:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.919651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.919701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.919715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.919736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:00 crc kubenswrapper[4715]: I1210 09:35:00.919752 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:00Z","lastTransitionTime":"2025-12-10T09:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.022470 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.022533 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.022551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.022576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.022590 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:01Z","lastTransitionTime":"2025-12-10T09:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.125737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.125777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.125786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.125801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.125810 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:01Z","lastTransitionTime":"2025-12-10T09:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.228207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.228501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.228581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.228659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.228723 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:01Z","lastTransitionTime":"2025-12-10T09:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.330820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.331100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.331204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.331313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.331394 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:01Z","lastTransitionTime":"2025-12-10T09:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.434020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.434050 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.434060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.434073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.434082 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:01Z","lastTransitionTime":"2025-12-10T09:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.536446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.536499 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.536509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.536524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.536534 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:01Z","lastTransitionTime":"2025-12-10T09:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.604673 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:01 crc kubenswrapper[4715]: E1210 09:35:01.604819 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.605097 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:01 crc kubenswrapper[4715]: E1210 09:35:01.605350 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.605583 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:01 crc kubenswrapper[4715]: E1210 09:35:01.605709 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.638837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.638875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.638884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.638900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.638911 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:01Z","lastTransitionTime":"2025-12-10T09:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.741505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.741548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.741560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.741579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.741590 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:01Z","lastTransitionTime":"2025-12-10T09:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.844014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.844077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.844097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.844121 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.844137 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:01Z","lastTransitionTime":"2025-12-10T09:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.946367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.946416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.946425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.946440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:01 crc kubenswrapper[4715]: I1210 09:35:01.946449 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:01Z","lastTransitionTime":"2025-12-10T09:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.049239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.049273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.049284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.049301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.049313 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:02Z","lastTransitionTime":"2025-12-10T09:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.151659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.151715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.151729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.151750 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.151762 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:02Z","lastTransitionTime":"2025-12-10T09:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.254866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.254928 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.254937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.254952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.254961 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:02Z","lastTransitionTime":"2025-12-10T09:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.357689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.357742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.357753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.357773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.357790 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:02Z","lastTransitionTime":"2025-12-10T09:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.413052 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:02 crc kubenswrapper[4715]: E1210 09:35:02.413264 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:35:02 crc kubenswrapper[4715]: E1210 09:35:02.413386 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs podName:f450d9cd-df1c-456e-861b-d91b2b683417 nodeName:}" failed. No retries permitted until 2025-12-10 09:35:34.413353799 +0000 UTC m=+97.156900090 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs") pod "network-metrics-daemon-h5xdb" (UID: "f450d9cd-df1c-456e-861b-d91b2b683417") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.460046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.460105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.460119 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.460137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.460155 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:02Z","lastTransitionTime":"2025-12-10T09:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.562424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.562469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.562479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.562495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.562504 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:02Z","lastTransitionTime":"2025-12-10T09:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.604064 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:02 crc kubenswrapper[4715]: E1210 09:35:02.604203 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.664892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.664959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.664970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.664984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.664993 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:02Z","lastTransitionTime":"2025-12-10T09:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.767891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.767937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.767946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.767960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.767969 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:02Z","lastTransitionTime":"2025-12-10T09:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.869956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.869998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.870007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.870021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.870031 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:02Z","lastTransitionTime":"2025-12-10T09:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.972061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.972121 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.972134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.972156 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:02 crc kubenswrapper[4715]: I1210 09:35:02.972165 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:02Z","lastTransitionTime":"2025-12-10T09:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.074851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.074901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.074930 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.074950 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.074962 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:03Z","lastTransitionTime":"2025-12-10T09:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.177373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.177431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.177449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.177469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.177484 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:03Z","lastTransitionTime":"2025-12-10T09:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.279420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.279463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.279475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.279491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.279501 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:03Z","lastTransitionTime":"2025-12-10T09:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.381793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.381829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.381837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.381851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.381859 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:03Z","lastTransitionTime":"2025-12-10T09:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.483905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.483981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.483997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.484018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.484035 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:03Z","lastTransitionTime":"2025-12-10T09:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.586292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.586337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.586350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.586367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.586380 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:03Z","lastTransitionTime":"2025-12-10T09:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.604570 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.604619 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.604619 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:03 crc kubenswrapper[4715]: E1210 09:35:03.604684 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:03 crc kubenswrapper[4715]: E1210 09:35:03.604814 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:03 crc kubenswrapper[4715]: E1210 09:35:03.604843 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.689004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.689071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.689084 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.689103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.689118 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:03Z","lastTransitionTime":"2025-12-10T09:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.791219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.791251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.791261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.791277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.791287 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:03Z","lastTransitionTime":"2025-12-10T09:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.893761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.893796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.893807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.893823 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.893835 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:03Z","lastTransitionTime":"2025-12-10T09:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.996175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.996219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.996234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.996253 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:03 crc kubenswrapper[4715]: I1210 09:35:03.996266 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:03Z","lastTransitionTime":"2025-12-10T09:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.098154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.098371 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.098491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.098582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.098653 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:04Z","lastTransitionTime":"2025-12-10T09:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.201109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.201151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.201162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.201177 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.201188 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:04Z","lastTransitionTime":"2025-12-10T09:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.304115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.304197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.304219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.304249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.304272 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:04Z","lastTransitionTime":"2025-12-10T09:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.408294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.408342 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.408354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.408372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.408385 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:04Z","lastTransitionTime":"2025-12-10T09:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.510780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.510834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.510848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.510868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.510882 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:04Z","lastTransitionTime":"2025-12-10T09:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.604561 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:04 crc kubenswrapper[4715]: E1210 09:35:04.604851 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.614398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.614448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.614457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.614477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.614488 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:04Z","lastTransitionTime":"2025-12-10T09:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.717493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.717539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.717548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.717564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.717573 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:04Z","lastTransitionTime":"2025-12-10T09:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.820423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.820467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.820476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.820495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.820505 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:04Z","lastTransitionTime":"2025-12-10T09:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.923004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.923049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.923065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.923086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:04 crc kubenswrapper[4715]: I1210 09:35:04.923101 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:04Z","lastTransitionTime":"2025-12-10T09:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.025869 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.025899 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.025907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.025935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.025943 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:05Z","lastTransitionTime":"2025-12-10T09:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.127774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.127907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.127965 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.127987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.128002 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:05Z","lastTransitionTime":"2025-12-10T09:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.230386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.230437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.230448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.230466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.230480 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:05Z","lastTransitionTime":"2025-12-10T09:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.333393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.333441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.333450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.333465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.333477 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:05Z","lastTransitionTime":"2025-12-10T09:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.435528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.435566 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.435576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.435754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.435775 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:05Z","lastTransitionTime":"2025-12-10T09:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.539297 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.539370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.539393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.539423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.539445 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:05Z","lastTransitionTime":"2025-12-10T09:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.604477 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:05 crc kubenswrapper[4715]: E1210 09:35:05.604702 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.605221 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:05 crc kubenswrapper[4715]: E1210 09:35:05.605448 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.605496 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:05 crc kubenswrapper[4715]: E1210 09:35:05.606148 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.606740 4715 scope.go:117] "RemoveContainer" containerID="1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e" Dec 10 09:35:05 crc kubenswrapper[4715]: E1210 09:35:05.607073 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.641547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.641594 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.641603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.641618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.641628 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:05Z","lastTransitionTime":"2025-12-10T09:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.744113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.744168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.744186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.744208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.744225 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:05Z","lastTransitionTime":"2025-12-10T09:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.846416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.846453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.846465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.846480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.846491 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:05Z","lastTransitionTime":"2025-12-10T09:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.948416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.948487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.948505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.948530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:05 crc kubenswrapper[4715]: I1210 09:35:05.948548 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:05Z","lastTransitionTime":"2025-12-10T09:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.051823 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.051872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.051881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.051899 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.051925 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:06Z","lastTransitionTime":"2025-12-10T09:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.154225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.154268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.154279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.154294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.154304 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:06Z","lastTransitionTime":"2025-12-10T09:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.257144 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.257222 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.257248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.257281 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.257306 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:06Z","lastTransitionTime":"2025-12-10T09:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.359784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.359827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.359838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.359853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.359862 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:06Z","lastTransitionTime":"2025-12-10T09:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.461791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.461856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.461866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.461879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.461888 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:06Z","lastTransitionTime":"2025-12-10T09:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.563965 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.563999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.564006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.564020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.564028 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:06Z","lastTransitionTime":"2025-12-10T09:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.604876 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:06 crc kubenswrapper[4715]: E1210 09:35:06.605229 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.613840 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.666951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.667021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.667033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.667059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.667075 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:06Z","lastTransitionTime":"2025-12-10T09:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.769574 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.769633 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.769651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.769674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.769689 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:06Z","lastTransitionTime":"2025-12-10T09:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.872437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.872492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.872504 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.872523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.872537 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:06Z","lastTransitionTime":"2025-12-10T09:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.975204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.975250 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.975261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.975278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:06 crc kubenswrapper[4715]: I1210 09:35:06.975291 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:06Z","lastTransitionTime":"2025-12-10T09:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.077275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.077326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.077339 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.077360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.077374 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:07Z","lastTransitionTime":"2025-12-10T09:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.180163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.180255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.180270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.180295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.180312 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:07Z","lastTransitionTime":"2025-12-10T09:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.277848 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/0.log" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.277957 4715 generic.go:334] "Generic (PLEG): container finished" podID="8870da50-36bc-413e-96ba-89740d93d5d9" containerID="09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855" exitCode=1 Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.278082 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-98zp6" event={"ID":"8870da50-36bc-413e-96ba-89740d93d5d9","Type":"ContainerDied","Data":"09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.279015 4715 scope.go:117] "RemoveContainer" containerID="09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.282839 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.282907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.282965 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.282987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.283001 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:07Z","lastTransitionTime":"2025-12-10T09:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.301056 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.319514 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.332979 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.349886 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.362309 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"834cd815-111b-4339-a87a-220afa018aba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4645809ac50eb70b0ad1ba89039994d876bc7114e3d40ba800e49d171f8f3fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.374889 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.385852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.385904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.385929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.385948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.385960 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:07Z","lastTransitionTime":"2025-12-10T09:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.388237 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.410792 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.423287 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.436283 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.452111 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.464549 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.483623 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:06Z\\\",\\\"message\\\":\\\"2025-12-10T09:34:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151\\\\n2025-12-10T09:34:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151 to /host/opt/cni/bin/\\\\n2025-12-10T09:34:21Z [verbose] multus-daemon started\\\\n2025-12-10T09:34:21Z [verbose] Readiness Indicator file check\\\\n2025-12-10T09:35:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.488592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.488649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.488661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.488681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.488693 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:07Z","lastTransitionTime":"2025-12-10T09:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.499790 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.513035 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.525115 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.536003 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.548881 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.558752 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.591614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.591653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.591663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.591682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.591695 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:07Z","lastTransitionTime":"2025-12-10T09:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.607482 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:07 crc kubenswrapper[4715]: E1210 09:35:07.607624 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.607717 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:07 crc kubenswrapper[4715]: E1210 09:35:07.607775 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.607825 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:07 crc kubenswrapper[4715]: E1210 09:35:07.607873 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.629830 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.644207 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.656542 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.671258 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.681138 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"834cd815-111b-4339-a87a-220afa018aba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4645809ac50eb70b0ad1ba89039994d876bc7114e3d40ba800e49d171f8f3fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.691842 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.694173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.694223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.694233 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.694251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.694260 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:07Z","lastTransitionTime":"2025-12-10T09:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.704278 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.724089 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.740331 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.792075 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.796940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.796995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.797007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.797030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.797043 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:07Z","lastTransitionTime":"2025-12-10T09:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.805266 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.816688 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.831360 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:06Z\\\",\\\"message\\\":\\\"2025-12-10T09:34:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151\\\\n2025-12-10T09:34:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151 to /host/opt/cni/bin/\\\\n2025-12-10T09:34:21Z [verbose] multus-daemon started\\\\n2025-12-10T09:34:21Z [verbose] Readiness Indicator file check\\\\n2025-12-10T09:35:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.906070 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.910477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.910522 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.910535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.910557 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.910568 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:07Z","lastTransitionTime":"2025-12-10T09:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.922030 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.945422 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.958089 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.967819 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:07 crc kubenswrapper[4715]: I1210 09:35:07.977268 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:07Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.013679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.013732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.013742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.014151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.014178 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:08Z","lastTransitionTime":"2025-12-10T09:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.117002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.117032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.117040 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.117054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.117063 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:08Z","lastTransitionTime":"2025-12-10T09:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.219352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.219395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.219403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.219418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.219429 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:08Z","lastTransitionTime":"2025-12-10T09:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.282992 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/0.log" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.283067 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-98zp6" event={"ID":"8870da50-36bc-413e-96ba-89740d93d5d9","Type":"ContainerStarted","Data":"3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b"} Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.293098 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"834cd815-111b-4339-a87a-220afa018aba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4645809ac50eb70b0ad1ba89039994d876bc7114e3d40ba800e49d171f8f3fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.304233 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.315076 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.321235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.321267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.321275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.321291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.321303 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:08Z","lastTransitionTime":"2025-12-10T09:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.337880 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.350758 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.361436 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.373988 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.385414 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.397986 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:06Z\\\",\\\"message\\\":\\\"2025-12-10T09:34:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151\\\\n2025-12-10T09:34:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151 to /host/opt/cni/bin/\\\\n2025-12-10T09:34:21Z [verbose] multus-daemon started\\\\n2025-12-10T09:34:21Z [verbose] Readiness Indicator file check\\\\n2025-12-10T09:35:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.409749 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.423270 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.423329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.423442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.423451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.423463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.423474 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:08Z","lastTransitionTime":"2025-12-10T09:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.443673 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.463018 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.475834 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.487755 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.508488 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.526128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.526206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.526233 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.526264 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.526290 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:08Z","lastTransitionTime":"2025-12-10T09:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.528534 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.541393 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.563489 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:08Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.604196 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:08 crc kubenswrapper[4715]: E1210 09:35:08.604575 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.628571 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.628660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.628684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.628717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.628739 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:08Z","lastTransitionTime":"2025-12-10T09:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.730729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.730770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.730779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.730793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.730806 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:08Z","lastTransitionTime":"2025-12-10T09:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.833874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.833939 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.833961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.833986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.834001 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:08Z","lastTransitionTime":"2025-12-10T09:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.936722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.936775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.936786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.936808 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:08 crc kubenswrapper[4715]: I1210 09:35:08.936821 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:08Z","lastTransitionTime":"2025-12-10T09:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.039772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.039835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.039847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.039866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.039877 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:09Z","lastTransitionTime":"2025-12-10T09:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.142180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.142228 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.142240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.142261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.142275 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:09Z","lastTransitionTime":"2025-12-10T09:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.244142 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.244189 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.244200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.244215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.244227 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:09Z","lastTransitionTime":"2025-12-10T09:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.346632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.346666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.346674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.346688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.346697 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:09Z","lastTransitionTime":"2025-12-10T09:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.449464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.449520 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.449533 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.449555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.449570 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:09Z","lastTransitionTime":"2025-12-10T09:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.552891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.553000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.553019 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.553044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.553060 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:09Z","lastTransitionTime":"2025-12-10T09:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.604944 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.604992 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.605043 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:09 crc kubenswrapper[4715]: E1210 09:35:09.605110 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:09 crc kubenswrapper[4715]: E1210 09:35:09.605346 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:09 crc kubenswrapper[4715]: E1210 09:35:09.605457 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.655663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.655732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.655746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.655770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.655786 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:09Z","lastTransitionTime":"2025-12-10T09:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.758059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.758103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.758112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.758125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.758133 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:09Z","lastTransitionTime":"2025-12-10T09:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.861044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.861126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.861150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.861180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.861198 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:09Z","lastTransitionTime":"2025-12-10T09:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.963886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.963986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.964006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.964033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:09 crc kubenswrapper[4715]: I1210 09:35:09.964053 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:09Z","lastTransitionTime":"2025-12-10T09:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.014624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.014691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.014707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.014729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.014741 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: E1210 09:35:10.028977 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:10Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.033651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.033726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.033749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.033775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.033793 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: E1210 09:35:10.046001 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:10Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.050157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.050196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.050209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.050227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.050238 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: E1210 09:35:10.062361 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:10Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.065962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.066047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.066077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.066110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.066135 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: E1210 09:35:10.081447 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:10Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.084718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.084769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.084780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.084801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.084813 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: E1210 09:35:10.096150 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:10Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:10 crc kubenswrapper[4715]: E1210 09:35:10.096264 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.097661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.097689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.097699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.097715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.097726 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.200494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.200576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.200588 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.200608 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.200619 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.302417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.302463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.302479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.302495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.302507 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.405390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.405438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.405451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.405470 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.405483 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.508798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.508859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.508878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.508904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.508953 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.604113 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:10 crc kubenswrapper[4715]: E1210 09:35:10.604270 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.611610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.611731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.611756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.611787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.611812 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.714622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.714722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.714738 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.714761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.714775 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.816769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.816851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.816874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.816903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.816960 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.920866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.920936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.920949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.920969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:10 crc kubenswrapper[4715]: I1210 09:35:10.920981 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:10Z","lastTransitionTime":"2025-12-10T09:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.023215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.023260 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.023270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.023288 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.023298 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:11Z","lastTransitionTime":"2025-12-10T09:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.126604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.126682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.126697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.126717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.126730 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:11Z","lastTransitionTime":"2025-12-10T09:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.229944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.229997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.230031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.230053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.230066 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:11Z","lastTransitionTime":"2025-12-10T09:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.333002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.333070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.333097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.333128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.333170 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:11Z","lastTransitionTime":"2025-12-10T09:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.435492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.435544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.435561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.435582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.435598 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:11Z","lastTransitionTime":"2025-12-10T09:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.538699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.538750 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.538767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.538789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.538806 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:11Z","lastTransitionTime":"2025-12-10T09:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.604529 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.604626 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.604533 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:11 crc kubenswrapper[4715]: E1210 09:35:11.604720 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:11 crc kubenswrapper[4715]: E1210 09:35:11.604805 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:11 crc kubenswrapper[4715]: E1210 09:35:11.605003 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.641335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.641377 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.641386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.641401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.641413 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:11Z","lastTransitionTime":"2025-12-10T09:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.744246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.744293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.744305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.744325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.744371 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:11Z","lastTransitionTime":"2025-12-10T09:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.943309 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.943338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.943347 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.943361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:11 crc kubenswrapper[4715]: I1210 09:35:11.943369 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:11Z","lastTransitionTime":"2025-12-10T09:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.045117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.045152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.045176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.045194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.045450 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:12Z","lastTransitionTime":"2025-12-10T09:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.148405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.148445 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.148455 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.148471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.148483 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:12Z","lastTransitionTime":"2025-12-10T09:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.250334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.250380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.250397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.250421 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.250436 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:12Z","lastTransitionTime":"2025-12-10T09:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.353280 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.353318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.353328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.353344 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.353356 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:12Z","lastTransitionTime":"2025-12-10T09:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.455424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.455464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.455475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.455491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.455502 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:12Z","lastTransitionTime":"2025-12-10T09:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.557471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.557525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.557537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.557556 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.557571 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:12Z","lastTransitionTime":"2025-12-10T09:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.604570 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:12 crc kubenswrapper[4715]: E1210 09:35:12.605069 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.660307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.660365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.660386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.660413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.660435 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:12Z","lastTransitionTime":"2025-12-10T09:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.763332 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.763401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.763425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.763462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.763488 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:12Z","lastTransitionTime":"2025-12-10T09:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.866173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.866261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.866315 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.866341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.866353 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:12Z","lastTransitionTime":"2025-12-10T09:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.967972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.968023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.968049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.968073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:12 crc kubenswrapper[4715]: I1210 09:35:12.968091 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:12Z","lastTransitionTime":"2025-12-10T09:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.071598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.071642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.071652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.071666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.071676 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:13Z","lastTransitionTime":"2025-12-10T09:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.175758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.176213 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.176231 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.176256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.176273 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:13Z","lastTransitionTime":"2025-12-10T09:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.279407 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.279478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.279495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.279519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.279537 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:13Z","lastTransitionTime":"2025-12-10T09:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.382405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.382446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.382460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.382481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.382496 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:13Z","lastTransitionTime":"2025-12-10T09:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.484929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.484979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.484994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.485010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.485018 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:13Z","lastTransitionTime":"2025-12-10T09:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.587377 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.587454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.587469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.587489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.587501 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:13Z","lastTransitionTime":"2025-12-10T09:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.605062 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.605160 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:13 crc kubenswrapper[4715]: E1210 09:35:13.605261 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.605159 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:13 crc kubenswrapper[4715]: E1210 09:35:13.605360 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:13 crc kubenswrapper[4715]: E1210 09:35:13.605615 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.690615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.690656 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.690665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.690681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.690691 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:13Z","lastTransitionTime":"2025-12-10T09:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.793664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.793721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.793732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.793751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.793763 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:13Z","lastTransitionTime":"2025-12-10T09:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.896776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.896842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.896860 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.896888 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.896906 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:13Z","lastTransitionTime":"2025-12-10T09:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:13 crc kubenswrapper[4715]: I1210 09:35:13.999815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:13.999973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.000002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.000035 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.000063 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:14Z","lastTransitionTime":"2025-12-10T09:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.102972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.103045 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.103072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.103102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.103125 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:14Z","lastTransitionTime":"2025-12-10T09:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.206584 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.206632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.206646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.206662 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.206673 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:14Z","lastTransitionTime":"2025-12-10T09:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.309310 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.309440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.309477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.309507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.309529 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:14Z","lastTransitionTime":"2025-12-10T09:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.412668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.412742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.412759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.412781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.412797 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:14Z","lastTransitionTime":"2025-12-10T09:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.515284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.515385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.515401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.515420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.515432 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:14Z","lastTransitionTime":"2025-12-10T09:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.604467 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:14 crc kubenswrapper[4715]: E1210 09:35:14.604634 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.617493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.617536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.617566 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.617590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.617607 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:14Z","lastTransitionTime":"2025-12-10T09:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.720412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.720450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.720460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.720475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.720486 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:14Z","lastTransitionTime":"2025-12-10T09:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.828752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.828819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.828877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.828903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.828955 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:14Z","lastTransitionTime":"2025-12-10T09:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.931726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.931787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.931804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.931829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:14 crc kubenswrapper[4715]: I1210 09:35:14.931846 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:14Z","lastTransitionTime":"2025-12-10T09:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.034875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.035008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.035027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.035053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.035070 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:15Z","lastTransitionTime":"2025-12-10T09:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.138081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.138136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.138154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.138179 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.138197 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:15Z","lastTransitionTime":"2025-12-10T09:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.241165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.241208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.241221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.241238 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.241252 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:15Z","lastTransitionTime":"2025-12-10T09:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.343734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.343798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.343816 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.343839 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.343856 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:15Z","lastTransitionTime":"2025-12-10T09:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.446080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.446134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.446151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.446172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.446189 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:15Z","lastTransitionTime":"2025-12-10T09:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.548146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.548225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.548251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.548286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.548310 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:15Z","lastTransitionTime":"2025-12-10T09:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.604225 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.604406 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:15 crc kubenswrapper[4715]: E1210 09:35:15.604497 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:15 crc kubenswrapper[4715]: E1210 09:35:15.604399 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.604225 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:15 crc kubenswrapper[4715]: E1210 09:35:15.604604 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.651000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.651066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.651091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.651121 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.651144 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:15Z","lastTransitionTime":"2025-12-10T09:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.754018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.754097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.754116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.754139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.754155 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:15Z","lastTransitionTime":"2025-12-10T09:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.857812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.857861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.857872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.857893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.857904 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:15Z","lastTransitionTime":"2025-12-10T09:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.961291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.961426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.961454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.961484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:15 crc kubenswrapper[4715]: I1210 09:35:15.961504 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:15Z","lastTransitionTime":"2025-12-10T09:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.063969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.064010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.064019 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.064032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.064041 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:16Z","lastTransitionTime":"2025-12-10T09:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.167009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.167068 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.167081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.167100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.167113 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:16Z","lastTransitionTime":"2025-12-10T09:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.269758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.269813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.269828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.269847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.269859 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:16Z","lastTransitionTime":"2025-12-10T09:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.373083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.373135 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.373146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.373163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.373173 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:16Z","lastTransitionTime":"2025-12-10T09:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.476800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.476859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.476870 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.476887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.476897 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:16Z","lastTransitionTime":"2025-12-10T09:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.579376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.579439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.579457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.579482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.579499 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:16Z","lastTransitionTime":"2025-12-10T09:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.604227 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:16 crc kubenswrapper[4715]: E1210 09:35:16.604933 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.605296 4715 scope.go:117] "RemoveContainer" containerID="1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.683407 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.683461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.683473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.683506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.683524 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:16Z","lastTransitionTime":"2025-12-10T09:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.786976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.787060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.787103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.787125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.787139 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:16Z","lastTransitionTime":"2025-12-10T09:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.890451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.890507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.890523 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.890545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.890564 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:16Z","lastTransitionTime":"2025-12-10T09:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.993896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.993972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.993986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.994005 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:16 crc kubenswrapper[4715]: I1210 09:35:16.994020 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:16Z","lastTransitionTime":"2025-12-10T09:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.096269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.096329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.096341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.096358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.096368 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:17Z","lastTransitionTime":"2025-12-10T09:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.198678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.198736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.198750 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.198771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.198785 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:17Z","lastTransitionTime":"2025-12-10T09:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.301660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.301690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.301698 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.301711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.301719 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:17Z","lastTransitionTime":"2025-12-10T09:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.319963 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/2.log" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.325827 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.326645 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103"} Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.327077 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.358516 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.387907 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.404021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.404069 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.404079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.404094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.404105 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:17Z","lastTransitionTime":"2025-12-10T09:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.408844 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.429190 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.440433 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.460793 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.470229 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"834cd815-111b-4339-a87a-220afa018aba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4645809ac50eb70b0ad1ba89039994d876bc7114e3d40ba800e49d171f8f3fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.490864 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.506731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.506974 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.507051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.507145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.507243 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:17Z","lastTransitionTime":"2025-12-10T09:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.509659 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.525529 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.535128 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.544119 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.553752 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:06Z\\\",\\\"message\\\":\\\"2025-12-10T09:34:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151\\\\n2025-12-10T09:34:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151 to /host/opt/cni/bin/\\\\n2025-12-10T09:34:21Z [verbose] multus-daemon started\\\\n2025-12-10T09:34:21Z [verbose] Readiness Indicator file check\\\\n2025-12-10T09:35:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.560972 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.568842 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.579793 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.591552 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.603821 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:17 crc kubenswrapper[4715]: E1210 09:35:17.604190 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.603897 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.603881 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.603874 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:17 crc kubenswrapper[4715]: E1210 09:35:17.604609 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:17 crc kubenswrapper[4715]: E1210 09:35:17.604622 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.610266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.610295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.610305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.610320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.610332 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:17Z","lastTransitionTime":"2025-12-10T09:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.619205 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.628405 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.644679 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"834cd815-111b-4339-a87a-220afa018aba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4645809ac50eb70b0ad1ba89039994d876bc7114e3d40ba800e49d171f8f3fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.658552 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.671440 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.695125 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.707898 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.712331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.712564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.712665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.712763 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.712840 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:17Z","lastTransitionTime":"2025-12-10T09:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.722885 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.739824 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.755085 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:06Z\\\",\\\"message\\\":\\\"2025-12-10T09:34:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151\\\\n2025-12-10T09:34:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151 to /host/opt/cni/bin/\\\\n2025-12-10T09:34:21Z [verbose] multus-daemon started\\\\n2025-12-10T09:34:21Z [verbose] Readiness Indicator file check\\\\n2025-12-10T09:35:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.766670 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.781163 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.795544 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.815983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.816023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.816039 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.816092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.816105 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:17Z","lastTransitionTime":"2025-12-10T09:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.816185 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.830930 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.842416 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.862387 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.878188 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.896357 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.915110 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:17Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.918613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.918651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.918659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.918675 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:17 crc kubenswrapper[4715]: I1210 09:35:17.918684 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:17Z","lastTransitionTime":"2025-12-10T09:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.026266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.026320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.026334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.026358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.026378 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:18Z","lastTransitionTime":"2025-12-10T09:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.129032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.129089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.129104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.129126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.129139 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:18Z","lastTransitionTime":"2025-12-10T09:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.232200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.232269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.232287 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.232310 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.232327 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:18Z","lastTransitionTime":"2025-12-10T09:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.336203 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.336263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.336273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.336292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.336305 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:18Z","lastTransitionTime":"2025-12-10T09:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.439509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.439567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.439586 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.439610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.439630 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:18Z","lastTransitionTime":"2025-12-10T09:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.543024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.543077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.543087 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.543102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.543111 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:18Z","lastTransitionTime":"2025-12-10T09:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.604765 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:18 crc kubenswrapper[4715]: E1210 09:35:18.610232 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.646197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.646248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.646261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.646281 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.646294 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:18Z","lastTransitionTime":"2025-12-10T09:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.748597 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.748651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.748668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.748690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.748703 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:18Z","lastTransitionTime":"2025-12-10T09:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.851558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.851592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.851600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.851615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.851632 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:18Z","lastTransitionTime":"2025-12-10T09:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.955597 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.955666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.955684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.955709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:18 crc kubenswrapper[4715]: I1210 09:35:18.955732 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:18Z","lastTransitionTime":"2025-12-10T09:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.058464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.058544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.058572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.058602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.058625 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:19Z","lastTransitionTime":"2025-12-10T09:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.161669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.161714 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.161726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.161744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.161764 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:19Z","lastTransitionTime":"2025-12-10T09:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.263847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.263890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.263903 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.263942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.263963 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:19Z","lastTransitionTime":"2025-12-10T09:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.335183 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/3.log" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.336339 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/2.log" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.340368 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.341488 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" exitCode=1 Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.341536 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.341577 4715 scope.go:117] "RemoveContainer" containerID="1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.342667 4715 scope.go:117] "RemoveContainer" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.342867 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.357475 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.366412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.366461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.366476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.366496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.366510 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:19Z","lastTransitionTime":"2025-12-10T09:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.372551 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.383841 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.399572 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.410810 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.421923 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.435500 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.453888 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.469436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.469465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.469473 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.469486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.469495 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:19Z","lastTransitionTime":"2025-12-10T09:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.469537 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.478825 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.494351 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:19Z\\\",\\\"message\\\":\\\"rnal_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-config-operator/metrics]} name:Service_openshift-config-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.161:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f32857b5-f652-4313-a0d7-455c3156dd99}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 09:35:18.106667 6727 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:18Z i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.503323 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.512188 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.521193 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"834cd815-111b-4339-a87a-220afa018aba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4645809ac50eb70b0ad1ba89039994d876bc7114e3d40ba800e49d171f8f3fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.531634 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.542193 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.555870 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.556550 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.556755 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:23.556682849 +0000 UTC m=+146.300229120 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.571347 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.574439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.574502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.574516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.574534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.574547 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:19Z","lastTransitionTime":"2025-12-10T09:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.582690 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:06Z\\\",\\\"message\\\":\\\"2025-12-10T09:34:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151\\\\n2025-12-10T09:34:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151 to /host/opt/cni/bin/\\\\n2025-12-10T09:34:21Z [verbose] multus-daemon started\\\\n2025-12-10T09:34:21Z [verbose] Readiness Indicator file check\\\\n2025-12-10T09:35:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:19Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.604257 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.604268 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.604389 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.604427 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.604514 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.604590 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.657991 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.658040 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.658061 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.658103 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658237 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658256 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658267 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658318 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 09:36:23.658300546 +0000 UTC m=+146.401846797 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658513 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658527 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658535 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658560 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 09:36:23.658551133 +0000 UTC m=+146.402097384 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658677 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658702 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:36:23.658694017 +0000 UTC m=+146.402240278 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658820 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:35:19 crc kubenswrapper[4715]: E1210 09:35:19.658846 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 09:36:23.658838911 +0000 UTC m=+146.402385162 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.677008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.677071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.677082 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.677099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.677110 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:19Z","lastTransitionTime":"2025-12-10T09:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.779804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.779862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.779878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.779900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.779951 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:19Z","lastTransitionTime":"2025-12-10T09:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.882818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.882896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.882931 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.882949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.882961 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:19Z","lastTransitionTime":"2025-12-10T09:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.986197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.986336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.986379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.986423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:19 crc kubenswrapper[4715]: I1210 09:35:19.986448 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:19Z","lastTransitionTime":"2025-12-10T09:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.089231 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.089276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.089294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.089315 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.089331 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.192605 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.192653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.192691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.192715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.192728 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.295682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.295939 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.296068 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.296177 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.296275 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.349485 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/3.log" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.350521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.350829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.351098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.351350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.351578 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.354483 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:35:20 crc kubenswrapper[4715]: E1210 09:35:20.376625 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.380876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.381154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.381239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.381325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.381405 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: E1210 09:35:20.393434 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.396498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.396532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.396545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.396564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.396579 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: E1210 09:35:20.407959 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.411990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.412026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.412038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.412056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.412069 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: E1210 09:35:20.424824 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.428580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.428687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.428750 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.428821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.428892 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: E1210 09:35:20.445533 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:20Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:20 crc kubenswrapper[4715]: E1210 09:35:20.445746 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.447336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.447384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.447396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.447417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.447430 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.549478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.549803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.550061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.550277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.550441 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.604616 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:20 crc kubenswrapper[4715]: E1210 09:35:20.605153 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.653742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.654168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.654301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.654452 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.654597 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.757545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.757943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.758103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.758260 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.758404 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.862098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.862555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.862797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.863158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.863395 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.967217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.967272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.967290 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.967315 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:20 crc kubenswrapper[4715]: I1210 09:35:20.967332 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:20Z","lastTransitionTime":"2025-12-10T09:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.071124 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.071215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.071242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.071335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.071364 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:21Z","lastTransitionTime":"2025-12-10T09:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.174109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.174185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.174209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.174241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.174264 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:21Z","lastTransitionTime":"2025-12-10T09:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.277899 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.278791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.279037 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.279263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.279512 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:21Z","lastTransitionTime":"2025-12-10T09:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.383636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.383721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.383746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.383839 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.383869 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:21Z","lastTransitionTime":"2025-12-10T09:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.486758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.486810 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.486822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.486846 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.486861 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:21Z","lastTransitionTime":"2025-12-10T09:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.590072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.590153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.590171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.590196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.590213 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:21Z","lastTransitionTime":"2025-12-10T09:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.604007 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.604007 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:21 crc kubenswrapper[4715]: E1210 09:35:21.604202 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:21 crc kubenswrapper[4715]: E1210 09:35:21.604364 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.604019 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:21 crc kubenswrapper[4715]: E1210 09:35:21.605176 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.692682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.692739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.692757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.692782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.692800 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:21Z","lastTransitionTime":"2025-12-10T09:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.795647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.795764 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.795797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.795830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.795851 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:21Z","lastTransitionTime":"2025-12-10T09:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.899214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.899284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.899305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.899348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:21 crc kubenswrapper[4715]: I1210 09:35:21.899378 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:21Z","lastTransitionTime":"2025-12-10T09:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.002079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.002150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.002173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.002204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.002226 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:22Z","lastTransitionTime":"2025-12-10T09:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.106122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.106181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.106217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.106245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.106266 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:22Z","lastTransitionTime":"2025-12-10T09:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.210090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.210149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.210171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.210201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.210221 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:22Z","lastTransitionTime":"2025-12-10T09:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.312729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.312793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.312811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.312828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.312841 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:22Z","lastTransitionTime":"2025-12-10T09:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.416455 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.416565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.416581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.416608 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.416624 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:22Z","lastTransitionTime":"2025-12-10T09:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.519285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.519340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.519358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.519378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.519392 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:22Z","lastTransitionTime":"2025-12-10T09:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.603832 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:22 crc kubenswrapper[4715]: E1210 09:35:22.604001 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.622079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.622141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.622162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.622185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.622210 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:22Z","lastTransitionTime":"2025-12-10T09:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.724349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.724385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.724398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.724411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.724419 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:22Z","lastTransitionTime":"2025-12-10T09:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.827340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.827395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.827404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.827418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.827427 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:22Z","lastTransitionTime":"2025-12-10T09:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.929389 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.929442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.929456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.929476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:22 crc kubenswrapper[4715]: I1210 09:35:22.929488 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:22Z","lastTransitionTime":"2025-12-10T09:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.033218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.033304 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.033329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.033362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.033387 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:23Z","lastTransitionTime":"2025-12-10T09:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.137062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.137118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.137184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.137212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.137227 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:23Z","lastTransitionTime":"2025-12-10T09:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.240959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.241012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.241024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.241042 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.241056 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:23Z","lastTransitionTime":"2025-12-10T09:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.343703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.343733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.343743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.343756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.343764 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:23Z","lastTransitionTime":"2025-12-10T09:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.446667 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.446735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.446753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.446787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.446815 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:23Z","lastTransitionTime":"2025-12-10T09:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.550167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.550211 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.550222 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.550238 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.550252 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:23Z","lastTransitionTime":"2025-12-10T09:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.604370 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.604465 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.604511 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:23 crc kubenswrapper[4715]: E1210 09:35:23.604686 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:23 crc kubenswrapper[4715]: E1210 09:35:23.604860 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:23 crc kubenswrapper[4715]: E1210 09:35:23.605018 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.653563 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.653712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.653732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.653759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.653776 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:23Z","lastTransitionTime":"2025-12-10T09:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.757996 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.758075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.758099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.758134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.758157 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:23Z","lastTransitionTime":"2025-12-10T09:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.861610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.861674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.861693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.861717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.861735 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:23Z","lastTransitionTime":"2025-12-10T09:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.964562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.964616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.964628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.964646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:23 crc kubenswrapper[4715]: I1210 09:35:23.964658 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:23Z","lastTransitionTime":"2025-12-10T09:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.067749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.067788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.067796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.067811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.067819 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:24Z","lastTransitionTime":"2025-12-10T09:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.170663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.170699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.170707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.170720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.170729 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:24Z","lastTransitionTime":"2025-12-10T09:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.274006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.274079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.274099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.274164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.274185 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:24Z","lastTransitionTime":"2025-12-10T09:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.377707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.377752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.377765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.377784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.377796 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:24Z","lastTransitionTime":"2025-12-10T09:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.480678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.480722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.480734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.480749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.480760 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:24Z","lastTransitionTime":"2025-12-10T09:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.583710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.583768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.583784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.583806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.583818 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:24Z","lastTransitionTime":"2025-12-10T09:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.604303 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:24 crc kubenswrapper[4715]: E1210 09:35:24.604466 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.686871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.686921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.686932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.686949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.686959 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:24Z","lastTransitionTime":"2025-12-10T09:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.789938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.789995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.790010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.790033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.790050 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:24Z","lastTransitionTime":"2025-12-10T09:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.893408 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.893492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.893530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.893562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.893586 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:24Z","lastTransitionTime":"2025-12-10T09:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.997239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.997290 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.997309 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.997331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:24 crc kubenswrapper[4715]: I1210 09:35:24.997352 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:24Z","lastTransitionTime":"2025-12-10T09:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.099564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.099641 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.099664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.099693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.099763 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:25Z","lastTransitionTime":"2025-12-10T09:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.203451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.203528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.203545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.203571 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.203588 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:25Z","lastTransitionTime":"2025-12-10T09:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.306757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.306807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.306822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.306841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.306857 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:25Z","lastTransitionTime":"2025-12-10T09:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.409025 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.409060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.409069 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.409082 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.409091 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:25Z","lastTransitionTime":"2025-12-10T09:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.511759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.511805 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.511816 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.511832 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.511843 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:25Z","lastTransitionTime":"2025-12-10T09:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.604956 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.605005 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.605039 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:25 crc kubenswrapper[4715]: E1210 09:35:25.605154 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:25 crc kubenswrapper[4715]: E1210 09:35:25.605520 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:25 crc kubenswrapper[4715]: E1210 09:35:25.605602 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.614094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.614324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.614417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.614509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.614596 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:25Z","lastTransitionTime":"2025-12-10T09:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.717295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.717352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.717369 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.717391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.717406 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:25Z","lastTransitionTime":"2025-12-10T09:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.819983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.820037 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.820046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.820059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.820067 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:25Z","lastTransitionTime":"2025-12-10T09:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.922879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.922925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.922933 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.922946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:25 crc kubenswrapper[4715]: I1210 09:35:25.922956 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:25Z","lastTransitionTime":"2025-12-10T09:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.025495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.025577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.025594 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.025616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.025631 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:26Z","lastTransitionTime":"2025-12-10T09:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.129242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.129277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.129303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.129318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.129326 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:26Z","lastTransitionTime":"2025-12-10T09:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.232610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.232652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.232660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.232674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.232683 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:26Z","lastTransitionTime":"2025-12-10T09:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.334727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.334771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.334788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.334807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.334818 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:26Z","lastTransitionTime":"2025-12-10T09:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.436795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.436843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.436854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.436871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.436883 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:26Z","lastTransitionTime":"2025-12-10T09:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.539051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.539120 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.539135 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.539153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.539169 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:26Z","lastTransitionTime":"2025-12-10T09:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.604954 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:26 crc kubenswrapper[4715]: E1210 09:35:26.605130 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.645392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.645451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.645501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.645525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.645538 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:26Z","lastTransitionTime":"2025-12-10T09:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.747541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.747572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.747580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.747592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.747601 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:26Z","lastTransitionTime":"2025-12-10T09:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.850248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.850298 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.850322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.850366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.850385 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:26Z","lastTransitionTime":"2025-12-10T09:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.953864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.953944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.953962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.953986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:26 crc kubenswrapper[4715]: I1210 09:35:26.954003 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:26Z","lastTransitionTime":"2025-12-10T09:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.056607 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.056721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.056753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.056784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.056806 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:27Z","lastTransitionTime":"2025-12-10T09:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.159199 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.159236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.159245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.159259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.159268 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:27Z","lastTransitionTime":"2025-12-10T09:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.261843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.261953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.261971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.261997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.262016 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:27Z","lastTransitionTime":"2025-12-10T09:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.365105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.365211 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.365236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.365270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.365291 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:27Z","lastTransitionTime":"2025-12-10T09:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.468228 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.468430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.468466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.468494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.468516 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:27Z","lastTransitionTime":"2025-12-10T09:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.571853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.571973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.571999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.572029 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.572053 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:27Z","lastTransitionTime":"2025-12-10T09:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.604745 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.604788 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.604795 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:27 crc kubenswrapper[4715]: E1210 09:35:27.605066 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:27 crc kubenswrapper[4715]: E1210 09:35:27.605131 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:27 crc kubenswrapper[4715]: E1210 09:35:27.605193 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.616684 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.628328 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.646511 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:06Z\\\",\\\"message\\\":\\\"2025-12-10T09:34:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151\\\\n2025-12-10T09:34:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151 to /host/opt/cni/bin/\\\\n2025-12-10T09:34:21Z [verbose] multus-daemon started\\\\n2025-12-10T09:34:21Z [verbose] Readiness Indicator file check\\\\n2025-12-10T09:35:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.662262 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.673650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.673724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.673749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.673780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.673803 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:27Z","lastTransitionTime":"2025-12-10T09:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.680322 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.691475 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.704424 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.716144 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.725989 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.749678 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.761457 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.772679 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.776067 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.776106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.776116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.776131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.776140 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:27Z","lastTransitionTime":"2025-12-10T09:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.788905 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.800134 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.811112 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"834cd815-111b-4339-a87a-220afa018aba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4645809ac50eb70b0ad1ba89039994d876bc7114e3d40ba800e49d171f8f3fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.823884 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.837200 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.859621 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a23d074d3eb8e710337f446160c118210d640583fba5853b75abe02bd57b76e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:49Z\\\",\\\"message\\\":\\\"otocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/control-plane-machine-set-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.41\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1210 09:34:49.653877 6378 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/packageserver-service for network=default are: map[]\\\\nF1210 09:34:49.654820 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:19Z\\\",\\\"message\\\":\\\"rnal_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-config-operator/metrics]} name:Service_openshift-config-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.161:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f32857b5-f652-4313-a0d7-455c3156dd99}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 09:35:18.106667 6727 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:18Z i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:35:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.873721 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:27Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.878053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.878110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.878123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.878139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.878150 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:27Z","lastTransitionTime":"2025-12-10T09:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.980462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.980511 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.980526 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.980544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:27 crc kubenswrapper[4715]: I1210 09:35:27.980558 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:27Z","lastTransitionTime":"2025-12-10T09:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.088938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.090555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.090584 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.090605 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.090619 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:28Z","lastTransitionTime":"2025-12-10T09:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.192461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.192501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.192512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.192528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.192540 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:28Z","lastTransitionTime":"2025-12-10T09:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.294884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.294937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.294954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.294970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.294980 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:28Z","lastTransitionTime":"2025-12-10T09:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.398010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.398085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.398112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.398146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.398168 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:28Z","lastTransitionTime":"2025-12-10T09:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.501294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.501367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.501387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.501409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.501427 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:28Z","lastTransitionTime":"2025-12-10T09:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.603908 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.604099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.604147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.604163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:28 crc kubenswrapper[4715]: E1210 09:35:28.604160 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.604186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.604275 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:28Z","lastTransitionTime":"2025-12-10T09:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.708203 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.708294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.708323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.708356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.708380 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:28Z","lastTransitionTime":"2025-12-10T09:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.812354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.812421 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.812438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.812465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.812483 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:28Z","lastTransitionTime":"2025-12-10T09:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.915852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.915941 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.915981 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.916012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:28 crc kubenswrapper[4715]: I1210 09:35:28.916038 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:28Z","lastTransitionTime":"2025-12-10T09:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.018927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.018978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.018990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.019006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.019019 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:29Z","lastTransitionTime":"2025-12-10T09:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.121697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.121737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.121749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.121770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.121782 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:29Z","lastTransitionTime":"2025-12-10T09:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.224944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.225000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.225019 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.225041 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.225054 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:29Z","lastTransitionTime":"2025-12-10T09:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.327248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.327284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.327292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.327306 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.327316 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:29Z","lastTransitionTime":"2025-12-10T09:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.429684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.429728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.429740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.429762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.429783 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:29Z","lastTransitionTime":"2025-12-10T09:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.532841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.532886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.532898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.532935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.532945 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:29Z","lastTransitionTime":"2025-12-10T09:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.604174 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.604319 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:29 crc kubenswrapper[4715]: E1210 09:35:29.604373 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.604523 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:29 crc kubenswrapper[4715]: E1210 09:35:29.604638 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:29 crc kubenswrapper[4715]: E1210 09:35:29.607787 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.635063 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.635129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.635153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.635178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.635193 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:29Z","lastTransitionTime":"2025-12-10T09:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.737889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.737969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.737989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.738011 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.738024 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:29Z","lastTransitionTime":"2025-12-10T09:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.840498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.840559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.840575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.840597 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.840612 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:29Z","lastTransitionTime":"2025-12-10T09:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.943765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.943848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.943869 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.943894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:29 crc kubenswrapper[4715]: I1210 09:35:29.943953 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:29Z","lastTransitionTime":"2025-12-10T09:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.046212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.046246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.046255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.046269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.046278 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.149218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.149289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.149314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.149349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.149372 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.251658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.251708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.251717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.251737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.251746 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.354056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.354100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.354111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.354128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.354140 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.456527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.456593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.456605 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.456623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.456634 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.558817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.558872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.558898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.558945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.558964 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.572652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.572718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.572735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.572758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.572772 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: E1210 09:35:30.590956 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.595299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.595367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.595386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.595405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.595417 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.604632 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:30 crc kubenswrapper[4715]: E1210 09:35:30.604773 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:30 crc kubenswrapper[4715]: E1210 09:35:30.607427 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.611269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.611305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.611316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.611332 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.611344 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: E1210 09:35:30.623677 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.627234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.627270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.627281 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.627295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.627307 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: E1210 09:35:30.639005 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.643601 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.643635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.643645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.643669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.643684 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: E1210 09:35:30.656005 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:30Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:30 crc kubenswrapper[4715]: E1210 09:35:30.656114 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.661984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.662038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.662051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.662072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.662085 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.765242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.765530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.765644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.765720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.765799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.869081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.869394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.869569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.869715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.869842 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.973199 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.973466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.973568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.973643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:30 crc kubenswrapper[4715]: I1210 09:35:30.973734 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:30Z","lastTransitionTime":"2025-12-10T09:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.077602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.078024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.078169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.078298 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.078412 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:31Z","lastTransitionTime":"2025-12-10T09:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.181153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.181187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.181198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.181214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.181225 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:31Z","lastTransitionTime":"2025-12-10T09:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.284817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.284892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.284945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.284980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.285003 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:31Z","lastTransitionTime":"2025-12-10T09:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.387655 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.387719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.387746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.387819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.387844 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:31Z","lastTransitionTime":"2025-12-10T09:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.490865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.490959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.490977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.491026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.491053 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:31Z","lastTransitionTime":"2025-12-10T09:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.593570 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.593617 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.593628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.593664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.593675 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:31Z","lastTransitionTime":"2025-12-10T09:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.604357 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:31 crc kubenswrapper[4715]: E1210 09:35:31.604493 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.604630 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.604640 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:31 crc kubenswrapper[4715]: E1210 09:35:31.605330 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:31 crc kubenswrapper[4715]: E1210 09:35:31.605557 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.605841 4715 scope.go:117] "RemoveContainer" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:35:31 crc kubenswrapper[4715]: E1210 09:35:31.606084 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.619632 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"834cd815-111b-4339-a87a-220afa018aba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4645809ac50eb70b0ad1ba89039994d876bc7114e3d40ba800e49d171f8f3fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.637062 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.651893 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.669258 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:19Z\\\",\\\"message\\\":\\\"rnal_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-config-operator/metrics]} name:Service_openshift-config-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.161:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f32857b5-f652-4313-a0d7-455c3156dd99}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 09:35:18.106667 6727 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:18Z i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:35:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.683624 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.693990 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.695276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.695318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.695331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.695347 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.695361 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:31Z","lastTransitionTime":"2025-12-10T09:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.707615 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.717718 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.728938 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:06Z\\\",\\\"message\\\":\\\"2025-12-10T09:34:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151\\\\n2025-12-10T09:34:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151 to /host/opt/cni/bin/\\\\n2025-12-10T09:34:21Z [verbose] multus-daemon started\\\\n2025-12-10T09:34:21Z [verbose] Readiness Indicator file check\\\\n2025-12-10T09:35:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.741117 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.753791 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.767178 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.782636 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.795663 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.798450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.798492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.798509 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.798532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.798548 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:31Z","lastTransitionTime":"2025-12-10T09:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.811264 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.833515 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.848696 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.859527 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.877181 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:31Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.901738 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.901791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.901804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.901829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:31 crc kubenswrapper[4715]: I1210 09:35:31.901842 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:31Z","lastTransitionTime":"2025-12-10T09:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.004576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.004610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.004618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.004632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.004648 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:32Z","lastTransitionTime":"2025-12-10T09:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.107463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.107524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.107539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.107560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.107578 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:32Z","lastTransitionTime":"2025-12-10T09:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.211115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.211192 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.211211 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.211246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.211268 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:32Z","lastTransitionTime":"2025-12-10T09:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.313975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.314032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.314047 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.314070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.314088 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:32Z","lastTransitionTime":"2025-12-10T09:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.417230 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.417286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.417303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.417326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.417343 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:32Z","lastTransitionTime":"2025-12-10T09:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.520767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.520838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.520847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.520867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.520878 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:32Z","lastTransitionTime":"2025-12-10T09:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.604804 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:32 crc kubenswrapper[4715]: E1210 09:35:32.605061 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.624139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.624182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.624194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.624218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.624234 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:32Z","lastTransitionTime":"2025-12-10T09:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.727268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.727324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.727338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.727358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.727371 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:32Z","lastTransitionTime":"2025-12-10T09:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.830454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.830497 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.830508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.830526 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.830537 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:32Z","lastTransitionTime":"2025-12-10T09:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.933221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.933472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.933482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.933495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:32 crc kubenswrapper[4715]: I1210 09:35:32.933504 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:32Z","lastTransitionTime":"2025-12-10T09:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.036098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.036134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.036149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.036170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.036184 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:33Z","lastTransitionTime":"2025-12-10T09:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.138625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.138685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.138705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.138728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.138746 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:33Z","lastTransitionTime":"2025-12-10T09:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.242312 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.242374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.242392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.242415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.242431 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:33Z","lastTransitionTime":"2025-12-10T09:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.345778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.345826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.345837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.345855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.345867 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:33Z","lastTransitionTime":"2025-12-10T09:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.447793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.447860 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.447880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.447905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.447952 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:33Z","lastTransitionTime":"2025-12-10T09:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.551331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.551381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.551393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.551407 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.551418 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:33Z","lastTransitionTime":"2025-12-10T09:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.603889 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.604058 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:33 crc kubenswrapper[4715]: E1210 09:35:33.604097 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.604136 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:33 crc kubenswrapper[4715]: E1210 09:35:33.604239 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:33 crc kubenswrapper[4715]: E1210 09:35:33.604405 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.654656 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.654758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.654785 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.654815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.654838 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:33Z","lastTransitionTime":"2025-12-10T09:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.758362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.758457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.758475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.758502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.758519 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:33Z","lastTransitionTime":"2025-12-10T09:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.861725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.861775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.861788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.861806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.861820 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:33Z","lastTransitionTime":"2025-12-10T09:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.964627 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.964714 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.964752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.964782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:33 crc kubenswrapper[4715]: I1210 09:35:33.964802 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:33Z","lastTransitionTime":"2025-12-10T09:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.068099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.068188 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.068208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.068237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.068257 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:34Z","lastTransitionTime":"2025-12-10T09:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.171756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.171830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.171849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.171876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.171894 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:34Z","lastTransitionTime":"2025-12-10T09:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.275906 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.276027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.276046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.276071 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.276088 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:34Z","lastTransitionTime":"2025-12-10T09:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.379343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.379404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.379423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.379446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.379487 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:34Z","lastTransitionTime":"2025-12-10T09:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:34 crc kubenswrapper[4715]: E1210 09:35:34.419083 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:35:34 crc kubenswrapper[4715]: E1210 09:35:34.419197 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs podName:f450d9cd-df1c-456e-861b-d91b2b683417 nodeName:}" failed. No retries permitted until 2025-12-10 09:36:38.419169332 +0000 UTC m=+161.162715623 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs") pod "network-metrics-daemon-h5xdb" (UID: "f450d9cd-df1c-456e-861b-d91b2b683417") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.418860 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.482439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.482512 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.482529 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.482554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.482571 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:34Z","lastTransitionTime":"2025-12-10T09:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.585994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.586038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.586049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.586069 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.586112 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:34Z","lastTransitionTime":"2025-12-10T09:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.604326 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:34 crc kubenswrapper[4715]: E1210 09:35:34.604478 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.689590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.689645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.689658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.689689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.689699 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:34Z","lastTransitionTime":"2025-12-10T09:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.792169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.792240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.792249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.792262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.792272 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:34Z","lastTransitionTime":"2025-12-10T09:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.894687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.894731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.894743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.894759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.894773 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:34Z","lastTransitionTime":"2025-12-10T09:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.997733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.997800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.997822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.997853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:34 crc kubenswrapper[4715]: I1210 09:35:34.997875 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:34Z","lastTransitionTime":"2025-12-10T09:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.100945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.100977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.100987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.101003 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.101013 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:35Z","lastTransitionTime":"2025-12-10T09:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.203726 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.203792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.203816 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.203845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.203862 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:35Z","lastTransitionTime":"2025-12-10T09:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.307072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.307131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.307142 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.307160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.307170 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:35Z","lastTransitionTime":"2025-12-10T09:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.409365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.409410 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.409420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.409435 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.409444 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:35Z","lastTransitionTime":"2025-12-10T09:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.512136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.512185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.512197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.512219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.512232 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:35Z","lastTransitionTime":"2025-12-10T09:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.603970 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.604033 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:35 crc kubenswrapper[4715]: E1210 09:35:35.604152 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.604198 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:35 crc kubenswrapper[4715]: E1210 09:35:35.604314 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:35 crc kubenswrapper[4715]: E1210 09:35:35.604347 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.615153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.615217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.615230 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.615248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.615262 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:35Z","lastTransitionTime":"2025-12-10T09:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.718639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.718719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.718733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.718753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.718781 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:35Z","lastTransitionTime":"2025-12-10T09:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.822219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.822276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.822295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.822319 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.822339 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:35Z","lastTransitionTime":"2025-12-10T09:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.925583 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.925645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.925660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.925678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:35 crc kubenswrapper[4715]: I1210 09:35:35.925690 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:35Z","lastTransitionTime":"2025-12-10T09:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.028992 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.029049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.029057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.029076 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.029088 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:36Z","lastTransitionTime":"2025-12-10T09:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.133678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.133731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.133745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.133767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.133778 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:36Z","lastTransitionTime":"2025-12-10T09:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.236335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.236381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.236390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.236407 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.236419 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:36Z","lastTransitionTime":"2025-12-10T09:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.338988 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.339048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.339059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.339081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.339093 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:36Z","lastTransitionTime":"2025-12-10T09:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.441539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.441590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.441602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.441619 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.441669 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:36Z","lastTransitionTime":"2025-12-10T09:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.544708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.544768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.544782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.544803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.544819 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:36Z","lastTransitionTime":"2025-12-10T09:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.604678 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:36 crc kubenswrapper[4715]: E1210 09:35:36.604854 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.646691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.646725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.646736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.646751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.646764 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:36Z","lastTransitionTime":"2025-12-10T09:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.749683 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.749743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.749755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.749776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.749787 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:36Z","lastTransitionTime":"2025-12-10T09:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.852471 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.852534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.852545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.852560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.852572 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:36Z","lastTransitionTime":"2025-12-10T09:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.955359 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.955403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.955416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.955434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:36 crc kubenswrapper[4715]: I1210 09:35:36.955446 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:36Z","lastTransitionTime":"2025-12-10T09:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.058432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.058496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.058519 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.058543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.058561 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:37Z","lastTransitionTime":"2025-12-10T09:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.161102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.161165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.161183 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.161208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.161223 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:37Z","lastTransitionTime":"2025-12-10T09:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.263366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.263441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.263464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.263495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.263518 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:37Z","lastTransitionTime":"2025-12-10T09:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.366380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.366439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.366454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.366476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.366495 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:37Z","lastTransitionTime":"2025-12-10T09:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.468909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.468964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.468976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.468994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.469005 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:37Z","lastTransitionTime":"2025-12-10T09:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.572680 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.572748 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.572767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.572794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.572814 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:37Z","lastTransitionTime":"2025-12-10T09:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.605018 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.605122 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:37 crc kubenswrapper[4715]: E1210 09:35:37.605236 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.605275 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:37 crc kubenswrapper[4715]: E1210 09:35:37.605362 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:37 crc kubenswrapper[4715]: E1210 09:35:37.605602 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.617877 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f450d9cd-df1c-456e-861b-d91b2b683417\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b5b5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-h5xdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.630035 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"834cd815-111b-4339-a87a-220afa018aba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4645809ac50eb70b0ad1ba89039994d876bc7114e3d40ba800e49d171f8f3fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://510a89e0933801d9186e289cff20e79b23b1a366c17ae613566dc8c93cf8a63a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.648046 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.663240 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d18a51c8f183efc3217abb87de284d822eeddb80e07faf4e68c968bba24b379f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.675538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.675600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.675615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.675636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.675652 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:37Z","lastTransitionTime":"2025-12-10T09:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.683553 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a44160c-cf30-4b13-b82c-4c402d967dd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:19Z\\\",\\\"message\\\":\\\"rnal_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-config-operator/metrics]} name:Service_openshift-config-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.161:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f32857b5-f652-4313-a0d7-455c3156dd99}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 09:35:18.106667 6727 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:18Z i\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:35:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sbxh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vspkd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.698066 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc3c82fe-6594-41d0-96d5-d8977ac76b15\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b06105f172f415a251d0b828086bfaedf8d8775ede8244e72a0e758fad0676\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca26d862ba08862737fe08e0adfb3515f73224bce6526e541295f8b6acdff7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hclwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rnj27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.710301 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36227d8b-137c-47d9-a593-c46d9a092210\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://262b53f349f3bc7010a831d763751b84eea5d3e33ae4ccb26cb1751ff75bf344\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6a4c28122f46351778c6ab2a88aa76f50b19eefe53749b53e76c4d826f4190\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec5638082f5d2f9df2407f63591c4abd4aefed011a7756240441658a8cdac8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8df188d03503a398a8e75efb7cf7a0f41e04e32f219935532a40e6abed697bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.722266 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.738643 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-98zp6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8870da50-36bc-413e-96ba-89740d93d5d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T09:35:06Z\\\",\\\"message\\\":\\\"2025-12-10T09:34:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151\\\\n2025-12-10T09:34:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_510ce56f-2eeb-446c-b940-725dab7f2151 to /host/opt/cni/bin/\\\\n2025-12-10T09:34:21Z [verbose] multus-daemon started\\\\n2025-12-10T09:34:21Z [verbose] Readiness Indicator file check\\\\n2025-12-10T09:35:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:35:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xdtqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-98zp6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.752767 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zddqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57659947-16fe-444e-a0d9-bc825e0aee39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc0999310cf5a36a620042db390b26f530908767d8ab8e9867d4e44cbaa49fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4grj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zddqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.769582 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad92614a-879e-4f8e-8066-8b3c6c395ee8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 09:34:09.978579 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 09:34:09.980417 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3499085645/tls.crt::/tmp/serving-cert-3499085645/tls.key\\\\\\\"\\\\nI1210 09:34:15.362390 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 09:34:15.364327 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 09:34:15.364345 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 09:34:15.364370 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 09:34:15.364376 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 09:34:15.386023 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1210 09:34:15.386140 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386188 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 09:34:15.386233 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 09:34:15.386270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 09:34:15.386306 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 09:34:15.386344 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1210 09:34:15.386600 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1210 09:34:15.387661 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.778465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.778490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.778499 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.778514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.778525 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:37Z","lastTransitionTime":"2025-12-10T09:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.782161 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fd44a74-6090-458f-a5cc-bdc40230f288\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ea23037214c073e97737d5680bc8dae2b2fc56aba509f0c575147f604a121b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a5e8afe4351725d6898a65682651016fbf2a6a9c854774e6c50b42f1602db5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74eb5d5b9b00b53168e25f42f257bf9e429f8cb9ada4b52343c9cf648f2c4d1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.797458 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51d93284ef218efa50bfa09b612529f19fd2f45d987361f9231557116326612e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.812535 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9e130c241095b0556aa8e39cfedd779e2bf74773ee2ec10a6623a3590f233a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2b6e630d9ccbf35a9202b6b6d7105f6862b3fd7d1c27b68dee627e64901f7f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.826772 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jn8jk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1498fcdd-3a54-4845-b1bf-5f2557f6fd0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485eac78b434fa5c8d9ac13d7a6f8f5c9debc1fead64d9af742ba832793293de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-654kj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jn8jk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.847980 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b565a85a-f068-4727-9243-3f4f33faa0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a01e72592c1bdf2eeeec3e1079efa0baa2e8b42891df4c102631cdab988809cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9c266cc13eeab68b08b65ee0a514a61439c1a8d25eb329c16e0893127d973db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28b6dd9b9d240f28c232d2286b6d3cb38194928246c2b5c595b9ad4c91c3f565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1648388a37a3b2c4d8fbb9795e22eec44518f3d66f0ea8c0ed766c1cf18cc4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://061211493af1fe8305ab78af9a4e80e98a306874170bdaff5e71e12ac9808a30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1666790603b73091c795b52f1d0e2bf288634c7e7ed195ce4a146e68d7e61f03\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b32c3cffc475b598b32620e473a18f493f86d0d44c47847a1a34ba7a1ced328f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:58Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400b080c17e86cfdb31d091d332d83d7ba287bdce8d19c8e3d8b69172e3f2c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:33:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:33:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.862024 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.874880 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b26b009-53b8-4260-88a0-0f9f9553d676\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20711120b73a04cfe4b0e6d0f054b7cb64ede05a80d59663924fb6195b8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98m8k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7pbhk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.880732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.880780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.880802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.880831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.880852 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:37Z","lastTransitionTime":"2025-12-10T09:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.893499 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bnh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32230eb5-c6d6-45ff-82db-78e70b283fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T09:34:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://070695e5b9e96be332bdde71e5ac504652c6ae14a1dbd696a467881527bb5807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T09:34:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49e3beebebcbe47cb188d53261463e8e2082bdcef696a2438dc25c3d6ae0fb2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://462b33f920691048ff19021ed3a18300f31ee2f16b913f1582d590e2999a1741\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://740fdac3eee2043e528a6d52159b62f1988573f0488dc0528484fecf6b9a809b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18c14edf7d1fd1888519ddc924966d71b3ab4d5bafc045b63f12059247700a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5de4878d837f4d52201442d2082eb105a22e68c70c4b3f256ed0163822f94854\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://918e45eae779901f06f2c8702e13c768f0b23cae45c0f20788c9b5f630e21501\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T09:34:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T09:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hqt9k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T09:34:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bnh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:37Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.983476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.983534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.983551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.983577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:37 crc kubenswrapper[4715]: I1210 09:35:37.983594 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:37Z","lastTransitionTime":"2025-12-10T09:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.086883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.087022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.087041 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.087066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.087083 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:38Z","lastTransitionTime":"2025-12-10T09:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.190176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.190241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.190257 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.190282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.190300 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:38Z","lastTransitionTime":"2025-12-10T09:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.293877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.293977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.293994 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.294016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.294033 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:38Z","lastTransitionTime":"2025-12-10T09:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.397157 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.397266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.397284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.397352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.397370 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:38Z","lastTransitionTime":"2025-12-10T09:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.501247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.501315 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.501336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.501364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.501386 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:38Z","lastTransitionTime":"2025-12-10T09:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.604047 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:38 crc kubenswrapper[4715]: E1210 09:35:38.604239 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.604796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.604883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.604898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.604947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.604961 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:38Z","lastTransitionTime":"2025-12-10T09:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.708151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.708223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.708242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.708268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.708286 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:38Z","lastTransitionTime":"2025-12-10T09:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.811325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.811389 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.811406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.811428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.811447 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:38Z","lastTransitionTime":"2025-12-10T09:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.914177 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.914246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.914262 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.914290 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:38 crc kubenswrapper[4715]: I1210 09:35:38.914307 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:38Z","lastTransitionTime":"2025-12-10T09:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.016853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.016906 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.016943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.016969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.016987 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:39Z","lastTransitionTime":"2025-12-10T09:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.120187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.120243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.120255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.120277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.120294 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:39Z","lastTransitionTime":"2025-12-10T09:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.223659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.223721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.223731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.223752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.223772 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:39Z","lastTransitionTime":"2025-12-10T09:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.326580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.326630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.326642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.326661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.326674 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:39Z","lastTransitionTime":"2025-12-10T09:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.429247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.429314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.429333 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.429385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.429402 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:39Z","lastTransitionTime":"2025-12-10T09:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.531306 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.531383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.531405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.531447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.531470 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:39Z","lastTransitionTime":"2025-12-10T09:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.604551 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.604585 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:39 crc kubenswrapper[4715]: E1210 09:35:39.604670 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.605053 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:39 crc kubenswrapper[4715]: E1210 09:35:39.605144 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:39 crc kubenswrapper[4715]: E1210 09:35:39.605279 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.634338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.634422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.634448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.634477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.634516 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:39Z","lastTransitionTime":"2025-12-10T09:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.737209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.737250 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.737265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.737282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.737294 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:39Z","lastTransitionTime":"2025-12-10T09:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.840404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.840448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.840460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.840486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.840499 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:39Z","lastTransitionTime":"2025-12-10T09:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.946451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.946502 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.946511 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.946531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:39 crc kubenswrapper[4715]: I1210 09:35:39.946542 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:39Z","lastTransitionTime":"2025-12-10T09:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.049780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.049826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.049837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.049856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.049870 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:40Z","lastTransitionTime":"2025-12-10T09:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.153198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.153385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.153418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.153453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.153479 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:40Z","lastTransitionTime":"2025-12-10T09:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.256294 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.256340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.256355 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.256373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.256383 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:40Z","lastTransitionTime":"2025-12-10T09:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.358532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.358568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.358576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.358591 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.358599 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:40Z","lastTransitionTime":"2025-12-10T09:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.461683 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.461717 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.461728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.461746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.461758 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:40Z","lastTransitionTime":"2025-12-10T09:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.564297 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.564365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.564375 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.564391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.564401 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:40Z","lastTransitionTime":"2025-12-10T09:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.604228 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:40 crc kubenswrapper[4715]: E1210 09:35:40.604434 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.667416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.667476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.667487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.667505 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.667518 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:40Z","lastTransitionTime":"2025-12-10T09:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.769829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.769871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.769882 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.769896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.769905 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:40Z","lastTransitionTime":"2025-12-10T09:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.872791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.872833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.872842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.872857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.872866 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:40Z","lastTransitionTime":"2025-12-10T09:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.979423 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.979474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.979493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.979513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:40 crc kubenswrapper[4715]: I1210 09:35:40.979525 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:40Z","lastTransitionTime":"2025-12-10T09:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.053775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.053833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.053844 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.053861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.053871 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: E1210 09:35:41.068219 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:41Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.074345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.074396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.074440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.074458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.074468 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: E1210 09:35:41.088026 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:41Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.091501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.091536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.091544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.091558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.091567 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: E1210 09:35:41.103194 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:41Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.107125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.107163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.107173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.107187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.107196 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: E1210 09:35:41.119937 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:41Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.124575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.124611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.124624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.124640 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.124654 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: E1210 09:35:41.135690 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T09:35:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"29a4a427-b049-4b60-be94-f8a6c917cb86\\\",\\\"systemUUID\\\":\\\"13f37a17-0606-45c2-b39f-55bf2b8fcb59\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T09:35:41Z is after 2025-08-24T17:21:41Z" Dec 10 09:35:41 crc kubenswrapper[4715]: E1210 09:35:41.135970 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.137615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.137665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.137676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.137691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.137700 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.240304 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.240363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.240371 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.240387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.240397 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.344477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.344555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.344578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.344610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.344636 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.446983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.447044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.447064 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.447091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.447108 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.551363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.551424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.551438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.551458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.551473 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.604946 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.605112 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.605356 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:41 crc kubenswrapper[4715]: E1210 09:35:41.605403 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:41 crc kubenswrapper[4715]: E1210 09:35:41.605619 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:41 crc kubenswrapper[4715]: E1210 09:35:41.605761 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.654494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.654539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.654549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.654567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.654583 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.757411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.757459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.757484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.757508 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.757523 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.860761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.860811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.860824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.860845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.860858 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.964240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.964305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.964324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.964343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:41 crc kubenswrapper[4715]: I1210 09:35:41.964355 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:41Z","lastTransitionTime":"2025-12-10T09:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.069684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.069769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.069799 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.069816 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.069828 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:42Z","lastTransitionTime":"2025-12-10T09:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.171978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.172012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.172020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.172033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.172043 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:42Z","lastTransitionTime":"2025-12-10T09:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.274467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.274516 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.274528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.274544 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.274554 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:42Z","lastTransitionTime":"2025-12-10T09:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.376955 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.376999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.377010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.377026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.377037 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:42Z","lastTransitionTime":"2025-12-10T09:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.480990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.481061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.481075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.481102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.481119 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:42Z","lastTransitionTime":"2025-12-10T09:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.584434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.584492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.584507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.584527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.584541 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:42Z","lastTransitionTime":"2025-12-10T09:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.604785 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:42 crc kubenswrapper[4715]: E1210 09:35:42.604991 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.687184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.687256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.687293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.687329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.687354 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:42Z","lastTransitionTime":"2025-12-10T09:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.796939 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.796990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.797004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.797027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.797044 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:42Z","lastTransitionTime":"2025-12-10T09:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.899721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.899801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.899817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.899847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:42 crc kubenswrapper[4715]: I1210 09:35:42.899867 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:42Z","lastTransitionTime":"2025-12-10T09:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.003426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.003495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.003510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.003538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.003555 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:43Z","lastTransitionTime":"2025-12-10T09:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.107661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.107738 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.107748 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.107765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.107776 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:43Z","lastTransitionTime":"2025-12-10T09:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.210214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.210278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.210291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.210320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.210333 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:43Z","lastTransitionTime":"2025-12-10T09:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.313092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.313173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.313205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.313235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.313256 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:43Z","lastTransitionTime":"2025-12-10T09:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.416438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.416494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.416507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.416530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.416546 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:43Z","lastTransitionTime":"2025-12-10T09:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.520422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.520482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.520494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.520513 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.520528 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:43Z","lastTransitionTime":"2025-12-10T09:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.604839 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.604981 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:43 crc kubenswrapper[4715]: E1210 09:35:43.605043 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.605131 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:43 crc kubenswrapper[4715]: E1210 09:35:43.605166 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:43 crc kubenswrapper[4715]: E1210 09:35:43.605349 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.623373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.623406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.623415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.623429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.623441 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:43Z","lastTransitionTime":"2025-12-10T09:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.726682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.726843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.726868 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.726895 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.726955 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:43Z","lastTransitionTime":"2025-12-10T09:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.829364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.829424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.829441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.829466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.829482 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:43Z","lastTransitionTime":"2025-12-10T09:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.931282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.931352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.931362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.931380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:43 crc kubenswrapper[4715]: I1210 09:35:43.931391 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:43Z","lastTransitionTime":"2025-12-10T09:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.035003 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.035068 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.035087 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.035111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.035128 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:44Z","lastTransitionTime":"2025-12-10T09:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.138668 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.138739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.138756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.138784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.138803 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:44Z","lastTransitionTime":"2025-12-10T09:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.242540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.242600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.242623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.242653 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.242675 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:44Z","lastTransitionTime":"2025-12-10T09:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.345496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.345576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.345600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.345633 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.345652 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:44Z","lastTransitionTime":"2025-12-10T09:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.448628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.448673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.448683 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.448703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.448714 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:44Z","lastTransitionTime":"2025-12-10T09:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.551612 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.551670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.551694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.551724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.551748 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:44Z","lastTransitionTime":"2025-12-10T09:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.604458 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:44 crc kubenswrapper[4715]: E1210 09:35:44.604767 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.606411 4715 scope.go:117] "RemoveContainer" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:35:44 crc kubenswrapper[4715]: E1210 09:35:44.606695 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.655535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.655679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.655718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.655743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.655760 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:44Z","lastTransitionTime":"2025-12-10T09:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.758570 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.758613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.758622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.758641 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.758655 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:44Z","lastTransitionTime":"2025-12-10T09:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.861537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.861597 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.861614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.861637 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.861654 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:44Z","lastTransitionTime":"2025-12-10T09:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.964687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.964734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.964753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.964783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:44 crc kubenswrapper[4715]: I1210 09:35:44.964799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:44Z","lastTransitionTime":"2025-12-10T09:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.067670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.067773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.067815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.067851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.067878 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:45Z","lastTransitionTime":"2025-12-10T09:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.170585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.170651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.170669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.170690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.170704 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:45Z","lastTransitionTime":"2025-12-10T09:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.273448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.273503 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.273517 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.273532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.273542 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:45Z","lastTransitionTime":"2025-12-10T09:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.376273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.376332 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.376341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.376361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.376374 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:45Z","lastTransitionTime":"2025-12-10T09:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.480085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.480171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.480185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.480205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.480223 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:45Z","lastTransitionTime":"2025-12-10T09:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.583123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.583180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.583190 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.583215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.583227 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:45Z","lastTransitionTime":"2025-12-10T09:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.604450 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.604557 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.604833 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:45 crc kubenswrapper[4715]: E1210 09:35:45.605108 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:45 crc kubenswrapper[4715]: E1210 09:35:45.605275 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:45 crc kubenswrapper[4715]: E1210 09:35:45.605283 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.685809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.685849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.685858 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.685871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.685882 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:45Z","lastTransitionTime":"2025-12-10T09:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.789036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.789114 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.789132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.789166 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.789188 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:45Z","lastTransitionTime":"2025-12-10T09:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.892380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.892449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.892467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.892496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.892521 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:45Z","lastTransitionTime":"2025-12-10T09:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.996273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.996340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.996362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.996409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:45 crc kubenswrapper[4715]: I1210 09:35:45.996431 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:45Z","lastTransitionTime":"2025-12-10T09:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.098875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.099007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.099044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.099075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.099096 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:46Z","lastTransitionTime":"2025-12-10T09:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.202411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.202480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.202490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.202506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.202515 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:46Z","lastTransitionTime":"2025-12-10T09:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.307558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.307662 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.307702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.307737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.307792 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:46Z","lastTransitionTime":"2025-12-10T09:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.411860 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.412313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.412348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.412381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.412404 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:46Z","lastTransitionTime":"2025-12-10T09:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.516241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.516314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.516351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.516379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.516399 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:46Z","lastTransitionTime":"2025-12-10T09:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.603904 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:46 crc kubenswrapper[4715]: E1210 09:35:46.604111 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.618440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.618468 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.618479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.618499 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.618511 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:46Z","lastTransitionTime":"2025-12-10T09:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.722331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.722386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.722400 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.722422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.722435 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:46Z","lastTransitionTime":"2025-12-10T09:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.825295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.825361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.825380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.825406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.825423 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:46Z","lastTransitionTime":"2025-12-10T09:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.927335 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.927373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.927382 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.927397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:46 crc kubenswrapper[4715]: I1210 09:35:46.927405 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:46Z","lastTransitionTime":"2025-12-10T09:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.030618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.030659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.030671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.030690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.030701 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:47Z","lastTransitionTime":"2025-12-10T09:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.132945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.133004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.133015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.133032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.133045 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:47Z","lastTransitionTime":"2025-12-10T09:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.236125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.236184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.236207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.236233 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.236251 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:47Z","lastTransitionTime":"2025-12-10T09:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.338839 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.338876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.338884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.338898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.338908 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:47Z","lastTransitionTime":"2025-12-10T09:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.442480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.442507 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.442514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.442526 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.442534 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:47Z","lastTransitionTime":"2025-12-10T09:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.545066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.545120 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.545132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.545150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.545163 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:47Z","lastTransitionTime":"2025-12-10T09:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.604037 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.604097 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.604214 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:47 crc kubenswrapper[4715]: E1210 09:35:47.604262 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:47 crc kubenswrapper[4715]: E1210 09:35:47.604401 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:47 crc kubenswrapper[4715]: E1210 09:35:47.604636 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.643595 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=41.643561841 podStartE2EDuration="41.643561841s" podCreationTimestamp="2025-12-10 09:35:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.642619864 +0000 UTC m=+110.386166125" watchObservedRunningTime="2025-12-10 09:35:47.643561841 +0000 UTC m=+110.387108092" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.648477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.648521 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.648565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.648585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.648598 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:47Z","lastTransitionTime":"2025-12-10T09:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.714393 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rnj27" podStartSLOduration=90.714366858 podStartE2EDuration="1m30.714366858s" podCreationTimestamp="2025-12-10 09:34:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.714137071 +0000 UTC m=+110.457683362" watchObservedRunningTime="2025-12-10 09:35:47.714366858 +0000 UTC m=+110.457913109" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.745750 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=57.745723663 podStartE2EDuration="57.745723663s" podCreationTimestamp="2025-12-10 09:34:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.730746396 +0000 UTC m=+110.474292647" watchObservedRunningTime="2025-12-10 09:35:47.745723663 +0000 UTC m=+110.489269924" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.751034 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.751075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.751085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.751101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.751114 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:47Z","lastTransitionTime":"2025-12-10T09:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.762617 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-98zp6" podStartSLOduration=91.762603066 podStartE2EDuration="1m31.762603066s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.761834603 +0000 UTC m=+110.505380864" watchObservedRunningTime="2025-12-10 09:35:47.762603066 +0000 UTC m=+110.506149317" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.774883 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zddqm" podStartSLOduration=91.774858264 podStartE2EDuration="1m31.774858264s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.774335138 +0000 UTC m=+110.517881389" watchObservedRunningTime="2025-12-10 09:35:47.774858264 +0000 UTC m=+110.518404515" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.805571 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=92.80555128 podStartE2EDuration="1m32.80555128s" podCreationTimestamp="2025-12-10 09:34:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.787586465 +0000 UTC m=+110.531132716" watchObservedRunningTime="2025-12-10 09:35:47.80555128 +0000 UTC m=+110.549097531" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.821579 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=91.821556077 podStartE2EDuration="1m31.821556077s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.805750565 +0000 UTC m=+110.549296816" watchObservedRunningTime="2025-12-10 09:35:47.821556077 +0000 UTC m=+110.565102348" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.853538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.853589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.853604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.853625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.853638 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:47Z","lastTransitionTime":"2025-12-10T09:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.878899 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-jn8jk" podStartSLOduration=91.87887563 podStartE2EDuration="1m31.87887563s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.845461334 +0000 UTC m=+110.589007585" watchObservedRunningTime="2025-12-10 09:35:47.87887563 +0000 UTC m=+110.622421871" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.879345 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=88.879336573 podStartE2EDuration="1m28.879336573s" podCreationTimestamp="2025-12-10 09:34:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.877348935 +0000 UTC m=+110.620895186" watchObservedRunningTime="2025-12-10 09:35:47.879336573 +0000 UTC m=+110.622882834" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.909120 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podStartSLOduration=91.909101062 podStartE2EDuration="1m31.909101062s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.908540146 +0000 UTC m=+110.652086407" watchObservedRunningTime="2025-12-10 09:35:47.909101062 +0000 UTC m=+110.652647313" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.929402 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-bnh42" podStartSLOduration=91.929371894 podStartE2EDuration="1m31.929371894s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:47.929298462 +0000 UTC m=+110.672844743" watchObservedRunningTime="2025-12-10 09:35:47.929371894 +0000 UTC m=+110.672918145" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.956110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.956150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.956159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.956174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:47 crc kubenswrapper[4715]: I1210 09:35:47.956182 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:47Z","lastTransitionTime":"2025-12-10T09:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.059302 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.059348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.059358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.059377 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.059387 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:48Z","lastTransitionTime":"2025-12-10T09:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.161078 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.161131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.161144 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.161160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.161171 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:48Z","lastTransitionTime":"2025-12-10T09:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.263352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.263404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.263419 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.263439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.263451 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:48Z","lastTransitionTime":"2025-12-10T09:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.365541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.365571 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.365580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.365595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.365606 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:48Z","lastTransitionTime":"2025-12-10T09:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.467236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.467278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.467286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.467300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.467312 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:48Z","lastTransitionTime":"2025-12-10T09:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.569062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.569102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.569111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.569125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.569134 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:48Z","lastTransitionTime":"2025-12-10T09:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.604391 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:48 crc kubenswrapper[4715]: E1210 09:35:48.604793 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.671201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.671242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.671268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.671289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.671302 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:48Z","lastTransitionTime":"2025-12-10T09:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.773501 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.773549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.773562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.773581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.773597 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:48Z","lastTransitionTime":"2025-12-10T09:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.875967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.876041 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.876053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.876068 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.876079 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:48Z","lastTransitionTime":"2025-12-10T09:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.979219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.979273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.979337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.979361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:48 crc kubenswrapper[4715]: I1210 09:35:48.979379 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:48Z","lastTransitionTime":"2025-12-10T09:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.081892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.081954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.081973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.081990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.082001 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:49Z","lastTransitionTime":"2025-12-10T09:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.184496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.184530 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.184541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.184573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.184582 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:49Z","lastTransitionTime":"2025-12-10T09:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.287887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.287976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.287995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.288019 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.288037 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:49Z","lastTransitionTime":"2025-12-10T09:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.391384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.391444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.391463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.391486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.391503 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:49Z","lastTransitionTime":"2025-12-10T09:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.494305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.494390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.494415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.494447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.494470 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:49Z","lastTransitionTime":"2025-12-10T09:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.597805 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.597885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.597908 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.597973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.597999 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:49Z","lastTransitionTime":"2025-12-10T09:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.604416 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.604490 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.604996 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:49 crc kubenswrapper[4715]: E1210 09:35:49.605108 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:49 crc kubenswrapper[4715]: E1210 09:35:49.605269 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:49 crc kubenswrapper[4715]: E1210 09:35:49.605305 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.701207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.701267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.701289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.701313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.701329 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:49Z","lastTransitionTime":"2025-12-10T09:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.804890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.804967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.804979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.805003 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.805014 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:49Z","lastTransitionTime":"2025-12-10T09:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.908525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.908576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.908589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.908609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:49 crc kubenswrapper[4715]: I1210 09:35:49.908644 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:49Z","lastTransitionTime":"2025-12-10T09:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.011839 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.012002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.012032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.012069 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.012093 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:50Z","lastTransitionTime":"2025-12-10T09:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.114641 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.114711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.114729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.114751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.114767 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:50Z","lastTransitionTime":"2025-12-10T09:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.217694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.217759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.217776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.217807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.217826 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:50Z","lastTransitionTime":"2025-12-10T09:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.321382 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.321438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.321455 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.321482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.321502 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:50Z","lastTransitionTime":"2025-12-10T09:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.424650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.424728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.424746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.424775 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.424787 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:50Z","lastTransitionTime":"2025-12-10T09:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.527842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.528266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.528449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.528648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.528839 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:50Z","lastTransitionTime":"2025-12-10T09:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.603834 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:50 crc kubenswrapper[4715]: E1210 09:35:50.604026 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.631349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.631387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.631398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.631414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.631425 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:50Z","lastTransitionTime":"2025-12-10T09:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.733391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.733437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.733448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.733465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.733483 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:50Z","lastTransitionTime":"2025-12-10T09:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.835524 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.835572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.835590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.835608 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.835619 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:50Z","lastTransitionTime":"2025-12-10T09:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.938506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.938584 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.938612 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.938643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:50 crc kubenswrapper[4715]: I1210 09:35:50.938665 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:50Z","lastTransitionTime":"2025-12-10T09:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.042010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.042109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.042134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.042171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.042198 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:51Z","lastTransitionTime":"2025-12-10T09:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.145971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.146031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.146054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.146087 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.146113 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:51Z","lastTransitionTime":"2025-12-10T09:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.255116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.255156 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.255168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.255184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.255195 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:51Z","lastTransitionTime":"2025-12-10T09:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.358638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.358689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.358708 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.358732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.358751 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:51Z","lastTransitionTime":"2025-12-10T09:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.461404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.461457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.461472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.461491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.461505 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:51Z","lastTransitionTime":"2025-12-10T09:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.499553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.500091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.500122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.500149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.500169 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T09:35:51Z","lastTransitionTime":"2025-12-10T09:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.613581 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:51 crc kubenswrapper[4715]: E1210 09:35:51.613781 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.613600 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.614030 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:51 crc kubenswrapper[4715]: E1210 09:35:51.614141 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:51 crc kubenswrapper[4715]: E1210 09:35:51.614230 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.621905 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj"] Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.622334 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.625089 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.625194 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.625243 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.625350 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.714968 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6f731db7-eda9-490b-bb49-a46537359d98-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.715054 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6f731db7-eda9-490b-bb49-a46537359d98-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.715089 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f731db7-eda9-490b-bb49-a46537359d98-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.715124 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f731db7-eda9-490b-bb49-a46537359d98-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.715150 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6f731db7-eda9-490b-bb49-a46537359d98-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.815993 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6f731db7-eda9-490b-bb49-a46537359d98-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.816098 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6f731db7-eda9-490b-bb49-a46537359d98-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.816179 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6f731db7-eda9-490b-bb49-a46537359d98-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.816207 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f731db7-eda9-490b-bb49-a46537359d98-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.816244 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f731db7-eda9-490b-bb49-a46537359d98-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.816241 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6f731db7-eda9-490b-bb49-a46537359d98-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.816334 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6f731db7-eda9-490b-bb49-a46537359d98-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.818208 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6f731db7-eda9-490b-bb49-a46537359d98-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.825707 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f731db7-eda9-490b-bb49-a46537359d98-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.834217 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f731db7-eda9-490b-bb49-a46537359d98-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5j9bj\" (UID: \"6f731db7-eda9-490b-bb49-a46537359d98\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:51 crc kubenswrapper[4715]: I1210 09:35:51.938177 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" Dec 10 09:35:52 crc kubenswrapper[4715]: I1210 09:35:52.462731 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" event={"ID":"6f731db7-eda9-490b-bb49-a46537359d98","Type":"ContainerStarted","Data":"11f025460dcecdc69212d136a9a0e8a59a7c94023481af0efce73039255ab095"} Dec 10 09:35:52 crc kubenswrapper[4715]: I1210 09:35:52.463070 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" event={"ID":"6f731db7-eda9-490b-bb49-a46537359d98","Type":"ContainerStarted","Data":"74cece7c69136f88387361388a218f58c180aedc137dd0633261d6152153a707"} Dec 10 09:35:52 crc kubenswrapper[4715]: I1210 09:35:52.477065 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5j9bj" podStartSLOduration=96.477048887 podStartE2EDuration="1m36.477048887s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:35:52.475654566 +0000 UTC m=+115.219200817" watchObservedRunningTime="2025-12-10 09:35:52.477048887 +0000 UTC m=+115.220595138" Dec 10 09:35:52 crc kubenswrapper[4715]: I1210 09:35:52.604355 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:52 crc kubenswrapper[4715]: E1210 09:35:52.604465 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:53 crc kubenswrapper[4715]: I1210 09:35:53.468790 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/1.log" Dec 10 09:35:53 crc kubenswrapper[4715]: I1210 09:35:53.469207 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/0.log" Dec 10 09:35:53 crc kubenswrapper[4715]: I1210 09:35:53.469246 4715 generic.go:334] "Generic (PLEG): container finished" podID="8870da50-36bc-413e-96ba-89740d93d5d9" containerID="3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b" exitCode=1 Dec 10 09:35:53 crc kubenswrapper[4715]: I1210 09:35:53.469276 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-98zp6" event={"ID":"8870da50-36bc-413e-96ba-89740d93d5d9","Type":"ContainerDied","Data":"3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b"} Dec 10 09:35:53 crc kubenswrapper[4715]: I1210 09:35:53.469309 4715 scope.go:117] "RemoveContainer" containerID="09c81192e2f63d6564e618bff05ea8bbbcd57a6079cf647f7bb18ced7e587855" Dec 10 09:35:53 crc kubenswrapper[4715]: I1210 09:35:53.469785 4715 scope.go:117] "RemoveContainer" containerID="3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b" Dec 10 09:35:53 crc kubenswrapper[4715]: E1210 09:35:53.470034 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-98zp6_openshift-multus(8870da50-36bc-413e-96ba-89740d93d5d9)\"" pod="openshift-multus/multus-98zp6" podUID="8870da50-36bc-413e-96ba-89740d93d5d9" Dec 10 09:35:53 crc kubenswrapper[4715]: I1210 09:35:53.606176 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:53 crc kubenswrapper[4715]: I1210 09:35:53.606228 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:53 crc kubenswrapper[4715]: I1210 09:35:53.606288 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:53 crc kubenswrapper[4715]: E1210 09:35:53.606344 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:53 crc kubenswrapper[4715]: E1210 09:35:53.606416 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:53 crc kubenswrapper[4715]: E1210 09:35:53.606515 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:54 crc kubenswrapper[4715]: I1210 09:35:54.475052 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/1.log" Dec 10 09:35:54 crc kubenswrapper[4715]: I1210 09:35:54.604786 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:54 crc kubenswrapper[4715]: E1210 09:35:54.605027 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:55 crc kubenswrapper[4715]: I1210 09:35:55.604438 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:55 crc kubenswrapper[4715]: I1210 09:35:55.604504 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:55 crc kubenswrapper[4715]: I1210 09:35:55.604456 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:55 crc kubenswrapper[4715]: E1210 09:35:55.604632 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:55 crc kubenswrapper[4715]: E1210 09:35:55.604744 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:55 crc kubenswrapper[4715]: E1210 09:35:55.604897 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:56 crc kubenswrapper[4715]: I1210 09:35:56.604722 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:56 crc kubenswrapper[4715]: E1210 09:35:56.604869 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:57 crc kubenswrapper[4715]: I1210 09:35:57.604737 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:57 crc kubenswrapper[4715]: I1210 09:35:57.606049 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:57 crc kubenswrapper[4715]: E1210 09:35:57.606042 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:57 crc kubenswrapper[4715]: I1210 09:35:57.606114 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:57 crc kubenswrapper[4715]: E1210 09:35:57.606261 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:57 crc kubenswrapper[4715]: E1210 09:35:57.606329 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:35:57 crc kubenswrapper[4715]: I1210 09:35:57.607043 4715 scope.go:117] "RemoveContainer" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:35:57 crc kubenswrapper[4715]: E1210 09:35:57.607239 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vspkd_openshift-ovn-kubernetes(2a44160c-cf30-4b13-b82c-4c402d967dd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" Dec 10 09:35:57 crc kubenswrapper[4715]: E1210 09:35:57.646043 4715 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 10 09:35:57 crc kubenswrapper[4715]: E1210 09:35:57.734324 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 09:35:58 crc kubenswrapper[4715]: I1210 09:35:58.603994 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:35:58 crc kubenswrapper[4715]: E1210 09:35:58.604216 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:35:59 crc kubenswrapper[4715]: I1210 09:35:59.604366 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:35:59 crc kubenswrapper[4715]: I1210 09:35:59.604500 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:35:59 crc kubenswrapper[4715]: E1210 09:35:59.604560 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:35:59 crc kubenswrapper[4715]: I1210 09:35:59.604407 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:35:59 crc kubenswrapper[4715]: E1210 09:35:59.604772 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:35:59 crc kubenswrapper[4715]: E1210 09:35:59.604671 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:36:00 crc kubenswrapper[4715]: I1210 09:36:00.604774 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:00 crc kubenswrapper[4715]: E1210 09:36:00.605068 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:36:01 crc kubenswrapper[4715]: I1210 09:36:01.605063 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:01 crc kubenswrapper[4715]: I1210 09:36:01.605124 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:01 crc kubenswrapper[4715]: I1210 09:36:01.605086 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:01 crc kubenswrapper[4715]: E1210 09:36:01.605239 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:36:01 crc kubenswrapper[4715]: E1210 09:36:01.605452 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:36:01 crc kubenswrapper[4715]: E1210 09:36:01.605568 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:36:02 crc kubenswrapper[4715]: I1210 09:36:02.604690 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:02 crc kubenswrapper[4715]: E1210 09:36:02.604955 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:36:02 crc kubenswrapper[4715]: E1210 09:36:02.735755 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 09:36:03 crc kubenswrapper[4715]: I1210 09:36:03.604485 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:03 crc kubenswrapper[4715]: I1210 09:36:03.604518 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:03 crc kubenswrapper[4715]: I1210 09:36:03.604636 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:03 crc kubenswrapper[4715]: E1210 09:36:03.604750 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:36:03 crc kubenswrapper[4715]: E1210 09:36:03.605053 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:36:03 crc kubenswrapper[4715]: E1210 09:36:03.605171 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:36:04 crc kubenswrapper[4715]: I1210 09:36:04.604400 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:04 crc kubenswrapper[4715]: E1210 09:36:04.604583 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:36:05 crc kubenswrapper[4715]: I1210 09:36:05.604235 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:05 crc kubenswrapper[4715]: E1210 09:36:05.604390 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:36:05 crc kubenswrapper[4715]: I1210 09:36:05.604632 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:05 crc kubenswrapper[4715]: E1210 09:36:05.604698 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:36:05 crc kubenswrapper[4715]: I1210 09:36:05.604904 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:05 crc kubenswrapper[4715]: E1210 09:36:05.605027 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:36:06 crc kubenswrapper[4715]: I1210 09:36:06.604347 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:06 crc kubenswrapper[4715]: E1210 09:36:06.604516 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:36:07 crc kubenswrapper[4715]: I1210 09:36:07.604640 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:07 crc kubenswrapper[4715]: I1210 09:36:07.604730 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:07 crc kubenswrapper[4715]: E1210 09:36:07.605999 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:36:07 crc kubenswrapper[4715]: I1210 09:36:07.606070 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:07 crc kubenswrapper[4715]: E1210 09:36:07.606153 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:36:07 crc kubenswrapper[4715]: E1210 09:36:07.606217 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:36:07 crc kubenswrapper[4715]: E1210 09:36:07.736496 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 09:36:08 crc kubenswrapper[4715]: I1210 09:36:08.604116 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:08 crc kubenswrapper[4715]: E1210 09:36:08.604645 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:36:08 crc kubenswrapper[4715]: I1210 09:36:08.604851 4715 scope.go:117] "RemoveContainer" containerID="3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b" Dec 10 09:36:09 crc kubenswrapper[4715]: I1210 09:36:09.604566 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:09 crc kubenswrapper[4715]: I1210 09:36:09.604676 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:09 crc kubenswrapper[4715]: E1210 09:36:09.604740 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:36:09 crc kubenswrapper[4715]: I1210 09:36:09.604830 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:09 crc kubenswrapper[4715]: E1210 09:36:09.605050 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:36:09 crc kubenswrapper[4715]: E1210 09:36:09.605174 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:36:09 crc kubenswrapper[4715]: I1210 09:36:09.643303 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/1.log" Dec 10 09:36:09 crc kubenswrapper[4715]: I1210 09:36:09.643395 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-98zp6" event={"ID":"8870da50-36bc-413e-96ba-89740d93d5d9","Type":"ContainerStarted","Data":"9262d98d872c33b9676c35f7142bb4e8274eee8158ccc7e65af9ba5eeeb62c6f"} Dec 10 09:36:10 crc kubenswrapper[4715]: I1210 09:36:10.604908 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:10 crc kubenswrapper[4715]: E1210 09:36:10.605114 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:36:11 crc kubenswrapper[4715]: I1210 09:36:11.604722 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:11 crc kubenswrapper[4715]: I1210 09:36:11.604851 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:11 crc kubenswrapper[4715]: E1210 09:36:11.604892 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:36:11 crc kubenswrapper[4715]: E1210 09:36:11.605175 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:36:11 crc kubenswrapper[4715]: I1210 09:36:11.605224 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:11 crc kubenswrapper[4715]: E1210 09:36:11.605849 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:36:11 crc kubenswrapper[4715]: I1210 09:36:11.606240 4715 scope.go:117] "RemoveContainer" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:36:12 crc kubenswrapper[4715]: I1210 09:36:12.476772 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-h5xdb"] Dec 10 09:36:12 crc kubenswrapper[4715]: I1210 09:36:12.477222 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:12 crc kubenswrapper[4715]: E1210 09:36:12.477333 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:36:12 crc kubenswrapper[4715]: I1210 09:36:12.656648 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/3.log" Dec 10 09:36:12 crc kubenswrapper[4715]: I1210 09:36:12.659764 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:36:12 crc kubenswrapper[4715]: I1210 09:36:12.660790 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerStarted","Data":"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f"} Dec 10 09:36:12 crc kubenswrapper[4715]: I1210 09:36:12.661321 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:36:12 crc kubenswrapper[4715]: I1210 09:36:12.696153 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podStartSLOduration=116.696135507 podStartE2EDuration="1m56.696135507s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:12.695464898 +0000 UTC m=+135.439011149" watchObservedRunningTime="2025-12-10 09:36:12.696135507 +0000 UTC m=+135.439681758" Dec 10 09:36:12 crc kubenswrapper[4715]: E1210 09:36:12.738452 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 09:36:13 crc kubenswrapper[4715]: I1210 09:36:13.604470 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:13 crc kubenswrapper[4715]: I1210 09:36:13.604596 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:13 crc kubenswrapper[4715]: E1210 09:36:13.604702 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:36:13 crc kubenswrapper[4715]: E1210 09:36:13.604762 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:36:13 crc kubenswrapper[4715]: I1210 09:36:13.605213 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:13 crc kubenswrapper[4715]: E1210 09:36:13.605416 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:36:14 crc kubenswrapper[4715]: I1210 09:36:14.604708 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:14 crc kubenswrapper[4715]: E1210 09:36:14.604946 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:36:15 crc kubenswrapper[4715]: I1210 09:36:15.603862 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:15 crc kubenswrapper[4715]: I1210 09:36:15.603893 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:15 crc kubenswrapper[4715]: E1210 09:36:15.604058 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:36:15 crc kubenswrapper[4715]: I1210 09:36:15.604157 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:15 crc kubenswrapper[4715]: E1210 09:36:15.604227 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:36:15 crc kubenswrapper[4715]: E1210 09:36:15.604351 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:36:16 crc kubenswrapper[4715]: I1210 09:36:16.604483 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:16 crc kubenswrapper[4715]: E1210 09:36:16.604701 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-h5xdb" podUID="f450d9cd-df1c-456e-861b-d91b2b683417" Dec 10 09:36:17 crc kubenswrapper[4715]: I1210 09:36:17.606950 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:17 crc kubenswrapper[4715]: I1210 09:36:17.606909 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:17 crc kubenswrapper[4715]: I1210 09:36:17.607023 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:17 crc kubenswrapper[4715]: E1210 09:36:17.609853 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 09:36:17 crc kubenswrapper[4715]: E1210 09:36:17.610111 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 09:36:17 crc kubenswrapper[4715]: E1210 09:36:17.610207 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 09:36:17 crc kubenswrapper[4715]: I1210 09:36:17.714382 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:36:17 crc kubenswrapper[4715]: I1210 09:36:17.714450 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:36:18 crc kubenswrapper[4715]: I1210 09:36:18.604559 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:18 crc kubenswrapper[4715]: I1210 09:36:18.607045 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 10 09:36:18 crc kubenswrapper[4715]: I1210 09:36:18.607119 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 10 09:36:19 crc kubenswrapper[4715]: I1210 09:36:19.604613 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:19 crc kubenswrapper[4715]: I1210 09:36:19.604613 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:19 crc kubenswrapper[4715]: I1210 09:36:19.604770 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:19 crc kubenswrapper[4715]: I1210 09:36:19.607222 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 10 09:36:19 crc kubenswrapper[4715]: I1210 09:36:19.607297 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 10 09:36:19 crc kubenswrapper[4715]: I1210 09:36:19.607296 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 10 09:36:19 crc kubenswrapper[4715]: I1210 09:36:19.607395 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 10 09:36:21 crc kubenswrapper[4715]: I1210 09:36:21.992253 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.032411 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.033021 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.033469 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-p5sqh"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.033991 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.034542 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-zphsm"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.035050 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.036043 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.036810 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.036984 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fd9v5"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.037307 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.037572 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-b9k84"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.037950 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-l7gt6"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.038374 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.038512 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.037575 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.038582 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.038739 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.037648 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.038996 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b9k84" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.039015 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.039565 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.039599 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.039649 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.039667 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.039729 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7qnsk"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.039898 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.039988 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.040173 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.042988 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.043729 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.043907 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.044016 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.044110 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.044337 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.051889 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.057297 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.057992 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.058502 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.058611 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.058718 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.070785 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.073154 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.073564 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.073990 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7lwft"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.074236 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.074451 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.075733 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.076928 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.077077 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.079241 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.079727 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.087637 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.087842 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.101765 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.102082 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.102741 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.102793 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.102902 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.103765 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.106002 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-j2l56"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.106953 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.106165 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.106255 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.106346 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.106405 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.106607 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.112762 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-x689f"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.113441 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.114053 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.115093 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.115402 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.115843 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.119517 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.119695 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.120012 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.120225 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.120392 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.120578 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.120777 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.120985 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.121256 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.121690 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.121952 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.122221 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.122487 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.122726 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.124207 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.124233 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.124776 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.124960 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.125006 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.124966 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.126422 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8wtv4"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.127169 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.127893 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.127964 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.128070 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.128197 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.128332 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.128444 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.128447 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.128757 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.128871 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.129267 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.129450 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.129559 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.129734 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.130481 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.130701 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.130798 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.130879 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.131116 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.131294 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.131425 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.131564 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.131718 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.132318 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.132437 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.132520 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.132615 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.132729 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.132675 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.133138 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.133305 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.133419 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.133542 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.133428 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.133680 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.133881 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.134016 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.134230 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.134413 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q5bbl"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.145050 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.145190 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.145338 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.145808 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.147093 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.153753 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-72qbm"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.155314 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.157221 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-q6645"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.166029 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.167972 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.168836 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.171974 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.172105 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nllws"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.172540 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.172945 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.173452 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.173629 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.173797 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.176623 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.177504 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.177681 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.176748 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.176797 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.180906 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.181446 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.186163 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xshb8"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.187429 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.188395 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.195009 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.195393 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.196397 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.196661 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.197796 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.198295 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.199184 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.199291 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fd9v5"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.199370 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b9k84"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.199449 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zphsm"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.199527 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.199697 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.200197 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.201990 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.202544 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.202716 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.203306 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.203491 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204440 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-images\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204472 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2gcc\" (UniqueName: \"kubernetes.io/projected/53adb33e-d25b-412a-be71-82e3e6380a3e-kube-api-access-z2gcc\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204494 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/834668e1-7a44-4844-b2f6-3b629ceeba22-config\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204512 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5669\" (UniqueName: \"kubernetes.io/projected/352662c6-4463-4850-89f1-e3aa2fbead2f-kube-api-access-d5669\") pod \"cluster-samples-operator-665b6dd947-p2967\" (UID: \"352662c6-4463-4850-89f1-e3aa2fbead2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204539 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54xwn\" (UniqueName: \"kubernetes.io/projected/834668e1-7a44-4844-b2f6-3b629ceeba22-kube-api-access-54xwn\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204558 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204577 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a7659e-76f5-49b6-abf8-c3facab139ab-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dgxcw\" (UID: \"52a7659e-76f5-49b6-abf8-c3facab139ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204596 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/834668e1-7a44-4844-b2f6-3b629ceeba22-service-ca-bundle\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204613 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204631 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1116225-4343-4dcf-b128-83a355de274e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nw2b\" (UID: \"a1116225-4343-4dcf-b128-83a355de274e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204648 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204666 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-service-ca\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204681 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm4gb\" (UniqueName: \"kubernetes.io/projected/f55a019d-460a-4b2e-b553-f8cd1166dd29-kube-api-access-tm4gb\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204699 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204715 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2ncb\" (UniqueName: \"kubernetes.io/projected/7d475c48-8e5c-48c3-9407-cce393aaca28-kube-api-access-c2ncb\") pod \"openshift-controller-manager-operator-756b6f6bc6-w2jc5\" (UID: \"7d475c48-8e5c-48c3-9407-cce393aaca28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204731 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk5kx\" (UniqueName: \"kubernetes.io/projected/5fbc3706-b2a3-48aa-a950-dc403907fa02-kube-api-access-jk5kx\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204748 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204764 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204780 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r26qr\" (UniqueName: \"kubernetes.io/projected/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-kube-api-access-r26qr\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204799 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-config\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204814 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be3075d3-3e88-4e8c-9aa7-41d7d5345290-config\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204833 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/352662c6-4463-4850-89f1-e3aa2fbead2f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p2967\" (UID: \"352662c6-4463-4850-89f1-e3aa2fbead2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204850 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a456b48-1823-4211-b1f9-cf64a5645e6a-trusted-ca\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204887 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5dfj\" (UniqueName: \"kubernetes.io/projected/a1116225-4343-4dcf-b128-83a355de274e-kube-api-access-q5dfj\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nw2b\" (UID: \"a1116225-4343-4dcf-b128-83a355de274e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204904 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204940 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/834668e1-7a44-4844-b2f6-3b629ceeba22-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204956 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204971 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be3075d3-3e88-4e8c-9aa7-41d7d5345290-trusted-ca\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.204989 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-image-import-ca\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205003 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnk2n\" (UniqueName: \"kubernetes.io/projected/be3075d3-3e88-4e8c-9aa7-41d7d5345290-kube-api-access-rnk2n\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205020 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-trusted-ca-bundle\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205043 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-etcd-serving-ca\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205059 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-oauth-config\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205074 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb8x4\" (UniqueName: \"kubernetes.io/projected/136c04d9-6eec-4412-91c3-0c426e0b7746-kube-api-access-gb8x4\") pod \"openshift-config-operator-7777fb866f-tsqzx\" (UID: \"136c04d9-6eec-4412-91c3-0c426e0b7746\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205089 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-audit\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205105 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205120 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205137 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/834668e1-7a44-4844-b2f6-3b629ceeba22-serving-cert\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205155 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a456b48-1823-4211-b1f9-cf64a5645e6a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205184 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f55a019d-460a-4b2e-b553-f8cd1166dd29-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205203 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52a7659e-76f5-49b6-abf8-c3facab139ab-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dgxcw\" (UID: \"52a7659e-76f5-49b6-abf8-c3facab139ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205221 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/53adb33e-d25b-412a-be71-82e3e6380a3e-etcd-client\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205241 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be3075d3-3e88-4e8c-9aa7-41d7d5345290-serving-cert\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205264 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f55a019d-460a-4b2e-b553-f8cd1166dd29-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205287 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d475c48-8e5c-48c3-9407-cce393aaca28-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-w2jc5\" (UID: \"7d475c48-8e5c-48c3-9407-cce393aaca28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205307 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/136c04d9-6eec-4412-91c3-0c426e0b7746-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tsqzx\" (UID: \"136c04d9-6eec-4412-91c3-0c426e0b7746\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205343 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/53adb33e-d25b-412a-be71-82e3e6380a3e-encryption-config\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205363 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53adb33e-d25b-412a-be71-82e3e6380a3e-audit-dir\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205382 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-oauth-serving-cert\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205403 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-policies\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205421 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vsn8\" (UniqueName: \"kubernetes.io/projected/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-kube-api-access-7vsn8\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205440 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53adb33e-d25b-412a-be71-82e3e6380a3e-serving-cert\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205461 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205482 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d475c48-8e5c-48c3-9407-cce393aaca28-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-w2jc5\" (UID: \"7d475c48-8e5c-48c3-9407-cce393aaca28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205500 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f55a019d-460a-4b2e-b553-f8cd1166dd29-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205535 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-serving-cert\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205582 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrzj2\" (UniqueName: \"kubernetes.io/projected/8a456b48-1823-4211-b1f9-cf64a5645e6a-kube-api-access-wrzj2\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205633 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205675 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/53adb33e-d25b-412a-be71-82e3e6380a3e-node-pullsecrets\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205696 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/136c04d9-6eec-4412-91c3-0c426e0b7746-serving-cert\") pod \"openshift-config-operator-7777fb866f-tsqzx\" (UID: \"136c04d9-6eec-4412-91c3-0c426e0b7746\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205716 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/39235b98-6c0c-4285-85f3-1137bc4c6fef-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lzmqn\" (UID: \"39235b98-6c0c-4285-85f3-1137bc4c6fef\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205737 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1116225-4343-4dcf-b128-83a355de274e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nw2b\" (UID: \"a1116225-4343-4dcf-b128-83a355de274e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205811 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-config\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.205869 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq8wt\" (UniqueName: \"kubernetes.io/projected/39235b98-6c0c-4285-85f3-1137bc4c6fef-kube-api-access-sq8wt\") pod \"package-server-manager-789f6589d5-lzmqn\" (UID: \"39235b98-6c0c-4285-85f3-1137bc4c6fef\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.206520 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nr6bd"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.209465 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.206547 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.210167 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjr6l\" (UniqueName: \"kubernetes.io/projected/52a7659e-76f5-49b6-abf8-c3facab139ab-kube-api-access-pjr6l\") pod \"openshift-apiserver-operator-796bbdcf4f-dgxcw\" (UID: \"52a7659e-76f5-49b6-abf8-c3facab139ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.210207 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9w26\" (UniqueName: \"kubernetes.io/projected/55cadec6-a68b-4d03-abc0-d9c628c3e4d0-kube-api-access-c9w26\") pod \"downloads-7954f5f757-b9k84\" (UID: \"55cadec6-a68b-4d03-abc0-d9c628c3e4d0\") " pod="openshift-console/downloads-7954f5f757-b9k84" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.210237 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-dir\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.210266 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-config\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.210287 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.210340 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a456b48-1823-4211-b1f9-cf64a5645e6a-metrics-tls\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.218414 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.218479 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.219252 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.219716 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.221207 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnj8g"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.221313 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.221364 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.327127 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.329210 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.330453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/834668e1-7a44-4844-b2f6-3b629ceeba22-serving-cert\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.330644 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a456b48-1823-4211-b1f9-cf64a5645e6a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.330765 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd0711ac-4869-4b85-b204-941a95b922d3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5h6sf\" (UID: \"cd0711ac-4869-4b85-b204-941a95b922d3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.330877 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f55a019d-460a-4b2e-b553-f8cd1166dd29-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.331012 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52a7659e-76f5-49b6-abf8-c3facab139ab-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dgxcw\" (UID: \"52a7659e-76f5-49b6-abf8-c3facab139ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.331142 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/53adb33e-d25b-412a-be71-82e3e6380a3e-etcd-client\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.331298 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be3075d3-3e88-4e8c-9aa7-41d7d5345290-serving-cert\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.331432 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/136c04d9-6eec-4412-91c3-0c426e0b7746-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tsqzx\" (UID: \"136c04d9-6eec-4412-91c3-0c426e0b7746\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.331544 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/53adb33e-d25b-412a-be71-82e3e6380a3e-encryption-config\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.331656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53adb33e-d25b-412a-be71-82e3e6380a3e-audit-dir\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.331783 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f55a019d-460a-4b2e-b553-f8cd1166dd29-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.331904 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d475c48-8e5c-48c3-9407-cce393aaca28-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-w2jc5\" (UID: \"7d475c48-8e5c-48c3-9407-cce393aaca28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332084 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-oauth-serving-cert\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332212 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-policies\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332339 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/412a7188-42b3-4f08-affa-f4df0c076e46-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-74xmx\" (UID: \"412a7188-42b3-4f08-affa-f4df0c076e46\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332470 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc7f7af1-1312-42c8-bc68-4bb42832b42d-service-ca-bundle\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332587 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vsn8\" (UniqueName: \"kubernetes.io/projected/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-kube-api-access-7vsn8\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332717 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53adb33e-d25b-412a-be71-82e3e6380a3e-serving-cert\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332847 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29764aad-227f-400b-9ebd-703c0e69f519-metrics-tls\") pod \"dns-operator-744455d44c-q5bbl\" (UID: \"29764aad-227f-400b-9ebd-703c0e69f519\") " pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332995 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c1b3ff-a655-44aa-9060-af44630df2a3-config\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.333125 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.333237 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d475c48-8e5c-48c3-9407-cce393aaca28-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-w2jc5\" (UID: \"7d475c48-8e5c-48c3-9407-cce393aaca28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.333348 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f55a019d-460a-4b2e-b553-f8cd1166dd29-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.333493 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.333639 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-serving-cert\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.333756 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrzj2\" (UniqueName: \"kubernetes.io/projected/8a456b48-1823-4211-b1f9-cf64a5645e6a-kube-api-access-wrzj2\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.333872 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1fb0006d-63be-4593-a64e-9af6225de71a-srv-cert\") pod \"catalog-operator-68c6474976-w76w7\" (UID: \"1fb0006d-63be-4593-a64e-9af6225de71a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.334025 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/53adb33e-d25b-412a-be71-82e3e6380a3e-node-pullsecrets\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.334152 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm6dw\" (UniqueName: \"kubernetes.io/projected/20ccb5ed-eda9-459d-9581-ad5b6d02cb9d-kube-api-access-vm6dw\") pod \"olm-operator-6b444d44fb-vt4cz\" (UID: \"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.334281 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/136c04d9-6eec-4412-91c3-0c426e0b7746-serving-cert\") pod \"openshift-config-operator-7777fb866f-tsqzx\" (UID: \"136c04d9-6eec-4412-91c3-0c426e0b7746\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.334401 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/39235b98-6c0c-4285-85f3-1137bc4c6fef-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lzmqn\" (UID: \"39235b98-6c0c-4285-85f3-1137bc4c6fef\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.334525 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cd0711ac-4869-4b85-b204-941a95b922d3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5h6sf\" (UID: \"cd0711ac-4869-4b85-b204-941a95b922d3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.334644 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/20ccb5ed-eda9-459d-9581-ad5b6d02cb9d-srv-cert\") pod \"olm-operator-6b444d44fb-vt4cz\" (UID: \"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.334779 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-config\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.334900 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq8wt\" (UniqueName: \"kubernetes.io/projected/39235b98-6c0c-4285-85f3-1137bc4c6fef-kube-api-access-sq8wt\") pod \"package-server-manager-789f6589d5-lzmqn\" (UID: \"39235b98-6c0c-4285-85f3-1137bc4c6fef\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.335152 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1116225-4343-4dcf-b128-83a355de274e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nw2b\" (UID: \"a1116225-4343-4dcf-b128-83a355de274e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.335285 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5tfz\" (UniqueName: \"kubernetes.io/projected/45bf2bc4-681f-46cf-aca3-93960c133428-kube-api-access-j5tfz\") pod \"control-plane-machine-set-operator-78cbb6b69f-tnrn5\" (UID: \"45bf2bc4-681f-46cf-aca3-93960c133428\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.335404 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9w26\" (UniqueName: \"kubernetes.io/projected/55cadec6-a68b-4d03-abc0-d9c628c3e4d0-kube-api-access-c9w26\") pod \"downloads-7954f5f757-b9k84\" (UID: \"55cadec6-a68b-4d03-abc0-d9c628c3e4d0\") " pod="openshift-console/downloads-7954f5f757-b9k84" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.335507 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-dir\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.335600 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.335707 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjr6l\" (UniqueName: \"kubernetes.io/projected/52a7659e-76f5-49b6-abf8-c3facab139ab-kube-api-access-pjr6l\") pod \"openshift-apiserver-operator-796bbdcf4f-dgxcw\" (UID: \"52a7659e-76f5-49b6-abf8-c3facab139ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.335789 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/45bf2bc4-681f-46cf-aca3-93960c133428-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-tnrn5\" (UID: \"45bf2bc4-681f-46cf-aca3-93960c133428\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.335880 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/412a7188-42b3-4f08-affa-f4df0c076e46-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-74xmx\" (UID: \"412a7188-42b3-4f08-affa-f4df0c076e46\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.335987 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97djd\" (UniqueName: \"kubernetes.io/projected/fc7f7af1-1312-42c8-bc68-4bb42832b42d-kube-api-access-97djd\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.336076 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-config\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.336154 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.336246 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a456b48-1823-4211-b1f9-cf64a5645e6a-metrics-tls\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.336331 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-images\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.336433 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2gcc\" (UniqueName: \"kubernetes.io/projected/53adb33e-d25b-412a-be71-82e3e6380a3e-kube-api-access-z2gcc\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.336607 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fc7f7af1-1312-42c8-bc68-4bb42832b42d-metrics-certs\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.336770 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/834668e1-7a44-4844-b2f6-3b629ceeba22-config\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.337032 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5669\" (UniqueName: \"kubernetes.io/projected/352662c6-4463-4850-89f1-e3aa2fbead2f-kube-api-access-d5669\") pod \"cluster-samples-operator-665b6dd947-p2967\" (UID: \"352662c6-4463-4850-89f1-e3aa2fbead2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.337188 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54xwn\" (UniqueName: \"kubernetes.io/projected/834668e1-7a44-4844-b2f6-3b629ceeba22-kube-api-access-54xwn\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.337327 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.337485 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0c1b3ff-a655-44aa-9060-af44630df2a3-auth-proxy-config\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.337613 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c0c1b3ff-a655-44aa-9060-af44630df2a3-machine-approver-tls\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.337737 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fc7f7af1-1312-42c8-bc68-4bb42832b42d-default-certificate\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.337868 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/834668e1-7a44-4844-b2f6-3b629ceeba22-service-ca-bundle\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.338017 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.338159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a7659e-76f5-49b6-abf8-c3facab139ab-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dgxcw\" (UID: \"52a7659e-76f5-49b6-abf8-c3facab139ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.338285 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1116225-4343-4dcf-b128-83a355de274e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nw2b\" (UID: \"a1116225-4343-4dcf-b128-83a355de274e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.338469 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/412a7188-42b3-4f08-affa-f4df0c076e46-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-74xmx\" (UID: \"412a7188-42b3-4f08-affa-f4df0c076e46\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.338593 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1fb0006d-63be-4593-a64e-9af6225de71a-profile-collector-cert\") pod \"catalog-operator-68c6474976-w76w7\" (UID: \"1fb0006d-63be-4593-a64e-9af6225de71a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.338710 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.338837 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0711ac-4869-4b85-b204-941a95b922d3-config\") pod \"kube-controller-manager-operator-78b949d7b-5h6sf\" (UID: \"cd0711ac-4869-4b85-b204-941a95b922d3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.338969 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/20ccb5ed-eda9-459d-9581-ad5b6d02cb9d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vt4cz\" (UID: \"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.339054 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d475c48-8e5c-48c3-9407-cce393aaca28-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-w2jc5\" (UID: \"7d475c48-8e5c-48c3-9407-cce393aaca28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.331994 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.339211 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eff08ce1-5a34-48f6-9f72-1e548f563b20-serving-cert\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.339334 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-service-ca\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.339542 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/53adb33e-d25b-412a-be71-82e3e6380a3e-encryption-config\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.339554 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm4gb\" (UniqueName: \"kubernetes.io/projected/f55a019d-460a-4b2e-b553-f8cd1166dd29-kube-api-access-tm4gb\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.339762 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/efb368e0-756b-4323-8ca2-0852ff385eb4-secret-volume\") pod \"collect-profiles-29422650-8r975\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.339877 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk5kx\" (UniqueName: \"kubernetes.io/projected/5fbc3706-b2a3-48aa-a950-dc403907fa02-kube-api-access-jk5kx\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.340014 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.340144 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-oauth-serving-cert\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.340154 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332141 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53adb33e-d25b-412a-be71-82e3e6380a3e-audit-dir\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.340441 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f55a019d-460a-4b2e-b553-f8cd1166dd29-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.340749 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-policies\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332610 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/136c04d9-6eec-4412-91c3-0c426e0b7746-available-featuregates\") pod \"openshift-config-operator-7777fb866f-tsqzx\" (UID: \"136c04d9-6eec-4412-91c3-0c426e0b7746\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332027 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.340954 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332117 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.341743 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2ncb\" (UniqueName: \"kubernetes.io/projected/7d475c48-8e5c-48c3-9407-cce393aaca28-kube-api-access-c2ncb\") pod \"openshift-controller-manager-operator-756b6f6bc6-w2jc5\" (UID: \"7d475c48-8e5c-48c3-9407-cce393aaca28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.336959 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be3075d3-3e88-4e8c-9aa7-41d7d5345290-serving-cert\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.336876 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/53adb33e-d25b-412a-be71-82e3e6380a3e-node-pullsecrets\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.339762 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/53adb33e-d25b-412a-be71-82e3e6380a3e-etcd-client\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332260 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.332318 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.339952 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/834668e1-7a44-4844-b2f6-3b629ceeba22-serving-cert\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344061 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344273 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-config\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344302 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344344 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r26qr\" (UniqueName: \"kubernetes.io/projected/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-kube-api-access-r26qr\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344366 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be3075d3-3e88-4e8c-9aa7-41d7d5345290-config\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344396 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwz7w\" (UniqueName: \"kubernetes.io/projected/eff08ce1-5a34-48f6-9f72-1e548f563b20-kube-api-access-gwz7w\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344422 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc72z\" (UniqueName: \"kubernetes.io/projected/efb368e0-756b-4323-8ca2-0852ff385eb4-kube-api-access-kc72z\") pod \"collect-profiles-29422650-8r975\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344444 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dclzs\" (UniqueName: \"kubernetes.io/projected/b06279e0-9b31-40fb-ac5b-d7fd85abd9c8-kube-api-access-dclzs\") pod \"multus-admission-controller-857f4d67dd-q6645\" (UID: \"b06279e0-9b31-40fb-ac5b-d7fd85abd9c8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344467 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zfqc\" (UniqueName: \"kubernetes.io/projected/4435aa00-3653-4733-90ef-86f45701c917-kube-api-access-6zfqc\") pod \"migrator-59844c95c7-8r4m5\" (UID: \"4435aa00-3653-4733-90ef-86f45701c917\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344489 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpw2t\" (UniqueName: \"kubernetes.io/projected/c0c1b3ff-a655-44aa-9060-af44630df2a3-kube-api-access-jpw2t\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.344543 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/352662c6-4463-4850-89f1-e3aa2fbead2f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p2967\" (UID: \"352662c6-4463-4850-89f1-e3aa2fbead2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.345171 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a456b48-1823-4211-b1f9-cf64a5645e6a-trusted-ca\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.345343 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efb368e0-756b-4323-8ca2-0852ff385eb4-config-volume\") pod \"collect-profiles-29422650-8r975\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.345560 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5dfj\" (UniqueName: \"kubernetes.io/projected/a1116225-4343-4dcf-b128-83a355de274e-kube-api-access-q5dfj\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nw2b\" (UID: \"a1116225-4343-4dcf-b128-83a355de274e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.345697 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fc7f7af1-1312-42c8-bc68-4bb42832b42d-stats-auth\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.345831 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.345975 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-config\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.345835 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53adb33e-d25b-412a-be71-82e3e6380a3e-serving-cert\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.345687 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-dir\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.346453 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-config\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.346500 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-config\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.345885 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1116225-4343-4dcf-b128-83a355de274e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nw2b\" (UID: \"a1116225-4343-4dcf-b128-83a355de274e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.347184 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/834668e1-7a44-4844-b2f6-3b629ceeba22-service-ca-bundle\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.347961 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-config\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.348614 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.348668 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f55a019d-460a-4b2e-b553-f8cd1166dd29-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.348809 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-images\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.349137 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.350040 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/834668e1-7a44-4844-b2f6-3b629ceeba22-config\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.350217 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a7659e-76f5-49b6-abf8-c3facab139ab-config\") pod \"openshift-apiserver-operator-796bbdcf4f-dgxcw\" (UID: \"52a7659e-76f5-49b6-abf8-c3facab139ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.350329 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/834668e1-7a44-4844-b2f6-3b629ceeba22-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.350680 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-service-ca\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.350772 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/39235b98-6c0c-4285-85f3-1137bc4c6fef-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-lzmqn\" (UID: \"39235b98-6c0c-4285-85f3-1137bc4c6fef\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.350796 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.350455 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.351590 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/352662c6-4463-4850-89f1-e3aa2fbead2f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p2967\" (UID: \"352662c6-4463-4850-89f1-e3aa2fbead2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.351671 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/834668e1-7a44-4844-b2f6-3b629ceeba22-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.351739 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52a7659e-76f5-49b6-abf8-c3facab139ab-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-dgxcw\" (UID: \"52a7659e-76f5-49b6-abf8-c3facab139ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.351778 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a456b48-1823-4211-b1f9-cf64a5645e6a-trusted-ca\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.351790 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be3075d3-3e88-4e8c-9aa7-41d7d5345290-config\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.351877 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be3075d3-3e88-4e8c-9aa7-41d7d5345290-trusted-ca\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352014 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6tsq\" (UniqueName: \"kubernetes.io/projected/1fb0006d-63be-4593-a64e-9af6225de71a-kube-api-access-b6tsq\") pod \"catalog-operator-68c6474976-w76w7\" (UID: \"1fb0006d-63be-4593-a64e-9af6225de71a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352201 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-trusted-ca-bundle\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352263 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-image-import-ca\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnk2n\" (UniqueName: \"kubernetes.io/projected/be3075d3-3e88-4e8c-9aa7-41d7d5345290-kube-api-access-rnk2n\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352342 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-client-ca\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352388 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrkkd\" (UniqueName: \"kubernetes.io/projected/29764aad-227f-400b-9ebd-703c0e69f519-kube-api-access-nrkkd\") pod \"dns-operator-744455d44c-q5bbl\" (UID: \"29764aad-227f-400b-9ebd-703c0e69f519\") " pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352401 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352441 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-etcd-serving-ca\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352532 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-oauth-config\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352564 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb8x4\" (UniqueName: \"kubernetes.io/projected/136c04d9-6eec-4412-91c3-0c426e0b7746-kube-api-access-gb8x4\") pod \"openshift-config-operator-7777fb866f-tsqzx\" (UID: \"136c04d9-6eec-4412-91c3-0c426e0b7746\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352592 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-audit\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352624 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352655 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352822 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b06279e0-9b31-40fb-ac5b-d7fd85abd9c8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-q6645\" (UID: \"b06279e0-9b31-40fb-ac5b-d7fd85abd9c8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.352996 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be3075d3-3e88-4e8c-9aa7-41d7d5345290-trusted-ca\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.353042 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xshb8"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.353276 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-trusted-ca-bundle\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.353794 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-etcd-serving-ca\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.354118 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.354155 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.354386 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.355703 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-audit\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.355869 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/53adb33e-d25b-412a-be71-82e3e6380a3e-image-import-ca\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.356045 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.356137 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.356407 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/136c04d9-6eec-4412-91c3-0c426e0b7746-serving-cert\") pod \"openshift-config-operator-7777fb866f-tsqzx\" (UID: \"136c04d9-6eec-4412-91c3-0c426e0b7746\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.356460 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.357035 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.357036 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.357309 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.357485 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1116225-4343-4dcf-b128-83a355de274e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nw2b\" (UID: \"a1116225-4343-4dcf-b128-83a355de274e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.358478 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.360350 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-serving-cert\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.357973 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d475c48-8e5c-48c3-9407-cce393aaca28-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-w2jc5\" (UID: \"7d475c48-8e5c-48c3-9407-cce393aaca28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.362408 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a456b48-1823-4211-b1f9-cf64a5645e6a-metrics-tls\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.362406 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-oauth-config\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.362554 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.364017 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.364085 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.365039 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.366034 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-p5sqh"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.367028 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-l7gt6"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.368034 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.368932 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7qnsk"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.370630 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-q6645"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.373853 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-j2l56"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.376810 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.377144 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-hg2rw"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.378796 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-m8hpb"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.378998 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.379969 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-m8hpb" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.379973 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.380979 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.382001 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nllws"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.382997 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q5bbl"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.384083 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.384969 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.385955 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.387879 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hg2rw"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.390903 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8wtv4"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.392004 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnj8g"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.393361 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.394787 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6njjb"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.396731 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7lwft"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.396892 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.397110 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-mm5xh"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.397869 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.398391 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6njjb"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.398482 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.399781 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.401253 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-m8hpb"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.402020 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.402755 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.403944 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.405119 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nr6bd"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.406279 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg"] Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.422514 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.436453 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454156 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/412a7188-42b3-4f08-affa-f4df0c076e46-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-74xmx\" (UID: \"412a7188-42b3-4f08-affa-f4df0c076e46\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454205 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1fb0006d-63be-4593-a64e-9af6225de71a-profile-collector-cert\") pod \"catalog-operator-68c6474976-w76w7\" (UID: \"1fb0006d-63be-4593-a64e-9af6225de71a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454231 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0711ac-4869-4b85-b204-941a95b922d3-config\") pod \"kube-controller-manager-operator-78b949d7b-5h6sf\" (UID: \"cd0711ac-4869-4b85-b204-941a95b922d3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454256 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/20ccb5ed-eda9-459d-9581-ad5b6d02cb9d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vt4cz\" (UID: \"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454280 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eff08ce1-5a34-48f6-9f72-1e548f563b20-serving-cert\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454312 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/efb368e0-756b-4323-8ca2-0852ff385eb4-secret-volume\") pod \"collect-profiles-29422650-8r975\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454360 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwz7w\" (UniqueName: \"kubernetes.io/projected/eff08ce1-5a34-48f6-9f72-1e548f563b20-kube-api-access-gwz7w\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454377 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc72z\" (UniqueName: \"kubernetes.io/projected/efb368e0-756b-4323-8ca2-0852ff385eb4-kube-api-access-kc72z\") pod \"collect-profiles-29422650-8r975\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454401 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dclzs\" (UniqueName: \"kubernetes.io/projected/b06279e0-9b31-40fb-ac5b-d7fd85abd9c8-kube-api-access-dclzs\") pod \"multus-admission-controller-857f4d67dd-q6645\" (UID: \"b06279e0-9b31-40fb-ac5b-d7fd85abd9c8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454421 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zfqc\" (UniqueName: \"kubernetes.io/projected/4435aa00-3653-4733-90ef-86f45701c917-kube-api-access-6zfqc\") pod \"migrator-59844c95c7-8r4m5\" (UID: \"4435aa00-3653-4733-90ef-86f45701c917\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454439 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpw2t\" (UniqueName: \"kubernetes.io/projected/c0c1b3ff-a655-44aa-9060-af44630df2a3-kube-api-access-jpw2t\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454476 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efb368e0-756b-4323-8ca2-0852ff385eb4-config-volume\") pod \"collect-profiles-29422650-8r975\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454507 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fc7f7af1-1312-42c8-bc68-4bb42832b42d-stats-auth\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454530 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-config\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454552 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6tsq\" (UniqueName: \"kubernetes.io/projected/1fb0006d-63be-4593-a64e-9af6225de71a-kube-api-access-b6tsq\") pod \"catalog-operator-68c6474976-w76w7\" (UID: \"1fb0006d-63be-4593-a64e-9af6225de71a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454590 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrkkd\" (UniqueName: \"kubernetes.io/projected/29764aad-227f-400b-9ebd-703c0e69f519-kube-api-access-nrkkd\") pod \"dns-operator-744455d44c-q5bbl\" (UID: \"29764aad-227f-400b-9ebd-703c0e69f519\") " pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454612 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-client-ca\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454655 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b06279e0-9b31-40fb-ac5b-d7fd85abd9c8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-q6645\" (UID: \"b06279e0-9b31-40fb-ac5b-d7fd85abd9c8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454683 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd0711ac-4869-4b85-b204-941a95b922d3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5h6sf\" (UID: \"cd0711ac-4869-4b85-b204-941a95b922d3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454707 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/412a7188-42b3-4f08-affa-f4df0c076e46-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-74xmx\" (UID: \"412a7188-42b3-4f08-affa-f4df0c076e46\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454726 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc7f7af1-1312-42c8-bc68-4bb42832b42d-service-ca-bundle\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454746 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29764aad-227f-400b-9ebd-703c0e69f519-metrics-tls\") pod \"dns-operator-744455d44c-q5bbl\" (UID: \"29764aad-227f-400b-9ebd-703c0e69f519\") " pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454767 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c1b3ff-a655-44aa-9060-af44630df2a3-config\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454791 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454820 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1fb0006d-63be-4593-a64e-9af6225de71a-srv-cert\") pod \"catalog-operator-68c6474976-w76w7\" (UID: \"1fb0006d-63be-4593-a64e-9af6225de71a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454845 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm6dw\" (UniqueName: \"kubernetes.io/projected/20ccb5ed-eda9-459d-9581-ad5b6d02cb9d-kube-api-access-vm6dw\") pod \"olm-operator-6b444d44fb-vt4cz\" (UID: \"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cd0711ac-4869-4b85-b204-941a95b922d3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5h6sf\" (UID: \"cd0711ac-4869-4b85-b204-941a95b922d3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454884 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/20ccb5ed-eda9-459d-9581-ad5b6d02cb9d-srv-cert\") pod \"olm-operator-6b444d44fb-vt4cz\" (UID: \"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454932 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5tfz\" (UniqueName: \"kubernetes.io/projected/45bf2bc4-681f-46cf-aca3-93960c133428-kube-api-access-j5tfz\") pod \"control-plane-machine-set-operator-78cbb6b69f-tnrn5\" (UID: \"45bf2bc4-681f-46cf-aca3-93960c133428\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454972 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/412a7188-42b3-4f08-affa-f4df0c076e46-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-74xmx\" (UID: \"412a7188-42b3-4f08-affa-f4df0c076e46\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.454994 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97djd\" (UniqueName: \"kubernetes.io/projected/fc7f7af1-1312-42c8-bc68-4bb42832b42d-kube-api-access-97djd\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.455021 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/45bf2bc4-681f-46cf-aca3-93960c133428-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-tnrn5\" (UID: \"45bf2bc4-681f-46cf-aca3-93960c133428\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.455073 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fc7f7af1-1312-42c8-bc68-4bb42832b42d-metrics-certs\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.455133 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0c1b3ff-a655-44aa-9060-af44630df2a3-auth-proxy-config\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.455156 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c0c1b3ff-a655-44aa-9060-af44630df2a3-machine-approver-tls\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.455175 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fc7f7af1-1312-42c8-bc68-4bb42832b42d-default-certificate\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.455943 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c1b3ff-a655-44aa-9060-af44630df2a3-config\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.455976 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-client-ca\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.456268 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0c1b3ff-a655-44aa-9060-af44630df2a3-auth-proxy-config\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.456346 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.457543 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-config\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.459847 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/efb368e0-756b-4323-8ca2-0852ff385eb4-secret-volume\") pod \"collect-profiles-29422650-8r975\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.460099 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eff08ce1-5a34-48f6-9f72-1e548f563b20-serving-cert\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.460210 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.460784 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/20ccb5ed-eda9-459d-9581-ad5b6d02cb9d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vt4cz\" (UID: \"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.461746 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c0c1b3ff-a655-44aa-9060-af44630df2a3-machine-approver-tls\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.463666 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/20ccb5ed-eda9-459d-9581-ad5b6d02cb9d-srv-cert\") pod \"olm-operator-6b444d44fb-vt4cz\" (UID: \"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.468414 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1fb0006d-63be-4593-a64e-9af6225de71a-profile-collector-cert\") pod \"catalog-operator-68c6474976-w76w7\" (UID: \"1fb0006d-63be-4593-a64e-9af6225de71a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.468530 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1fb0006d-63be-4593-a64e-9af6225de71a-srv-cert\") pod \"catalog-operator-68c6474976-w76w7\" (UID: \"1fb0006d-63be-4593-a64e-9af6225de71a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.475423 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.515328 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.535251 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.550324 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29764aad-227f-400b-9ebd-703c0e69f519-metrics-tls\") pod \"dns-operator-744455d44c-q5bbl\" (UID: \"29764aad-227f-400b-9ebd-703c0e69f519\") " pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.555952 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.575272 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.595119 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.599192 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fc7f7af1-1312-42c8-bc68-4bb42832b42d-default-certificate\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.616211 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.634796 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.642451 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fc7f7af1-1312-42c8-bc68-4bb42832b42d-stats-auth\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.655374 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.662901 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fc7f7af1-1312-42c8-bc68-4bb42832b42d-metrics-certs\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.675298 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.676462 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc7f7af1-1312-42c8-bc68-4bb42832b42d-service-ca-bundle\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.695511 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.708568 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b06279e0-9b31-40fb-ac5b-d7fd85abd9c8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-q6645\" (UID: \"b06279e0-9b31-40fb-ac5b-d7fd85abd9c8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.715843 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.735685 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.756499 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.770029 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/45bf2bc4-681f-46cf-aca3-93960c133428-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-tnrn5\" (UID: \"45bf2bc4-681f-46cf-aca3-93960c133428\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.774753 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.795526 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.815218 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.834666 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.856152 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.875709 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.894882 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.915518 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.936849 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.956686 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.979731 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 10 09:36:22 crc kubenswrapper[4715]: I1210 09:36:22.996552 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.015658 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.034971 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.056529 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.076192 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.102569 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.107703 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd0711ac-4869-4b85-b204-941a95b922d3-config\") pod \"kube-controller-manager-operator-78b949d7b-5h6sf\" (UID: \"cd0711ac-4869-4b85-b204-941a95b922d3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.115439 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.135995 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.142036 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd0711ac-4869-4b85-b204-941a95b922d3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5h6sf\" (UID: \"cd0711ac-4869-4b85-b204-941a95b922d3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.156167 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.175398 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.196221 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.214015 4715 request.go:700] Waited for 1.013858374s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-oauth-apiserver/secrets?fieldSelector=metadata.name%3Dserving-cert&limit=500&resourceVersion=0 Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.215688 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.236031 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.255691 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.276108 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.295564 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.315803 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.335727 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.355732 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.369440 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/412a7188-42b3-4f08-affa-f4df0c076e46-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-74xmx\" (UID: \"412a7188-42b3-4f08-affa-f4df0c076e46\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.375610 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.395203 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.405273 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/412a7188-42b3-4f08-affa-f4df0c076e46-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-74xmx\" (UID: \"412a7188-42b3-4f08-affa-f4df0c076e46\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.416058 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.426355 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efb368e0-756b-4323-8ca2-0852ff385eb4-config-volume\") pod \"collect-profiles-29422650-8r975\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.436089 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.456130 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.475228 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.495046 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.536080 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.556600 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.571644 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:23 crc kubenswrapper[4715]: E1210 09:36:23.571754 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:38:25.571728818 +0000 UTC m=+268.315275069 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.575362 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.595871 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.616185 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.636281 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.655891 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.673266 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.673302 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.673409 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.673539 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.674773 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.675704 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.676830 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.677165 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.677318 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.696134 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.715301 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.735588 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.754971 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.783197 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.796152 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.822047 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.829773 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.831337 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a456b48-1823-4211-b1f9-cf64a5645e6a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.836012 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.839524 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.870449 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f55a019d-460a-4b2e-b553-f8cd1166dd29-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.891106 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vsn8\" (UniqueName: \"kubernetes.io/projected/c9ce5159-4cdb-47c7-9193-8adde2d4e2d0-kube-api-access-7vsn8\") pod \"machine-api-operator-5694c8668f-fd9v5\" (UID: \"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.913589 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54xwn\" (UniqueName: \"kubernetes.io/projected/834668e1-7a44-4844-b2f6-3b629ceeba22-kube-api-access-54xwn\") pod \"authentication-operator-69f744f599-l7gt6\" (UID: \"834668e1-7a44-4844-b2f6-3b629ceeba22\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.934564 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq8wt\" (UniqueName: \"kubernetes.io/projected/39235b98-6c0c-4285-85f3-1137bc4c6fef-kube-api-access-sq8wt\") pod \"package-server-manager-789f6589d5-lzmqn\" (UID: \"39235b98-6c0c-4285-85f3-1137bc4c6fef\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.947792 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.952284 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2ncb\" (UniqueName: \"kubernetes.io/projected/7d475c48-8e5c-48c3-9407-cce393aaca28-kube-api-access-c2ncb\") pod \"openshift-controller-manager-operator-756b6f6bc6-w2jc5\" (UID: \"7d475c48-8e5c-48c3-9407-cce393aaca28\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:23 crc kubenswrapper[4715]: I1210 09:36:23.954792 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.107845 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.107860 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.113377 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9w26\" (UniqueName: \"kubernetes.io/projected/55cadec6-a68b-4d03-abc0-d9c628c3e4d0-kube-api-access-c9w26\") pod \"downloads-7954f5f757-b9k84\" (UID: \"55cadec6-a68b-4d03-abc0-d9c628c3e4d0\") " pod="openshift-console/downloads-7954f5f757-b9k84" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.148219 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5669\" (UniqueName: \"kubernetes.io/projected/352662c6-4463-4850-89f1-e3aa2fbead2f-kube-api-access-d5669\") pod \"cluster-samples-operator-665b6dd947-p2967\" (UID: \"352662c6-4463-4850-89f1-e3aa2fbead2f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.149604 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk5kx\" (UniqueName: \"kubernetes.io/projected/5fbc3706-b2a3-48aa-a950-dc403907fa02-kube-api-access-jk5kx\") pod \"console-f9d7485db-zphsm\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.172249 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjr6l\" (UniqueName: \"kubernetes.io/projected/52a7659e-76f5-49b6-abf8-c3facab139ab-kube-api-access-pjr6l\") pod \"openshift-apiserver-operator-796bbdcf4f-dgxcw\" (UID: \"52a7659e-76f5-49b6-abf8-c3facab139ab\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.173462 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.174048 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm4gb\" (UniqueName: \"kubernetes.io/projected/f55a019d-460a-4b2e-b553-f8cd1166dd29-kube-api-access-tm4gb\") pod \"cluster-image-registry-operator-dc59b4c8b-gvqxr\" (UID: \"f55a019d-460a-4b2e-b553-f8cd1166dd29\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.174117 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2gcc\" (UniqueName: \"kubernetes.io/projected/53adb33e-d25b-412a-be71-82e3e6380a3e-kube-api-access-z2gcc\") pod \"apiserver-76f77b778f-p5sqh\" (UID: \"53adb33e-d25b-412a-be71-82e3e6380a3e\") " pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.174121 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrzj2\" (UniqueName: \"kubernetes.io/projected/8a456b48-1823-4211-b1f9-cf64a5645e6a-kube-api-access-wrzj2\") pod \"ingress-operator-5b745b69d9-7bhpw\" (UID: \"8a456b48-1823-4211-b1f9-cf64a5645e6a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.175633 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5dfj\" (UniqueName: \"kubernetes.io/projected/a1116225-4343-4dcf-b128-83a355de274e-kube-api-access-q5dfj\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nw2b\" (UID: \"a1116225-4343-4dcf-b128-83a355de274e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.175657 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.188002 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnk2n\" (UniqueName: \"kubernetes.io/projected/be3075d3-3e88-4e8c-9aa7-41d7d5345290-kube-api-access-rnk2n\") pod \"console-operator-58897d9998-7lwft\" (UID: \"be3075d3-3e88-4e8c-9aa7-41d7d5345290\") " pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.216020 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.233788 4715 request.go:700] Waited for 1.853159722s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Dcanary-serving-cert&limit=500&resourceVersion=0 Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.235574 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.240197 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.240270 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.240291 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.240423 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r26qr\" (UniqueName: \"kubernetes.io/projected/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-kube-api-access-r26qr\") pod \"oauth-openshift-558db77b4-7qnsk\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.240588 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.240673 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.244162 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.244424 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.255061 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.282418 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.290545 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb8x4\" (UniqueName: \"kubernetes.io/projected/136c04d9-6eec-4412-91c3-0c426e0b7746-kube-api-access-gb8x4\") pod \"openshift-config-operator-7777fb866f-tsqzx\" (UID: \"136c04d9-6eec-4412-91c3-0c426e0b7746\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.350262 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.367469 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b9k84" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.368044 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.368113 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.374202 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.374211 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.374446 4715 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.375940 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.396867 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.415284 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.437722 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.460421 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwz7w\" (UniqueName: \"kubernetes.io/projected/eff08ce1-5a34-48f6-9f72-1e548f563b20-kube-api-access-gwz7w\") pod \"controller-manager-879f6c89f-j2l56\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.534786 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc72z\" (UniqueName: \"kubernetes.io/projected/efb368e0-756b-4323-8ca2-0852ff385eb4-kube-api-access-kc72z\") pod \"collect-profiles-29422650-8r975\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.614250 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.618112 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97djd\" (UniqueName: \"kubernetes.io/projected/fc7f7af1-1312-42c8-bc68-4bb42832b42d-kube-api-access-97djd\") pod \"router-default-5444994796-72qbm\" (UID: \"fc7f7af1-1312-42c8-bc68-4bb42832b42d\") " pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.632857 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/412a7188-42b3-4f08-affa-f4df0c076e46-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-74xmx\" (UID: \"412a7188-42b3-4f08-affa-f4df0c076e46\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.637877 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.682681 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dclzs\" (UniqueName: \"kubernetes.io/projected/b06279e0-9b31-40fb-ac5b-d7fd85abd9c8-kube-api-access-dclzs\") pod \"multus-admission-controller-857f4d67dd-q6645\" (UID: \"b06279e0-9b31-40fb-ac5b-d7fd85abd9c8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.687978 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zfqc\" (UniqueName: \"kubernetes.io/projected/4435aa00-3653-4733-90ef-86f45701c917-kube-api-access-6zfqc\") pod \"migrator-59844c95c7-8r4m5\" (UID: \"4435aa00-3653-4733-90ef-86f45701c917\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.690008 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5tfz\" (UniqueName: \"kubernetes.io/projected/45bf2bc4-681f-46cf-aca3-93960c133428-kube-api-access-j5tfz\") pod \"control-plane-machine-set-operator-78cbb6b69f-tnrn5\" (UID: \"45bf2bc4-681f-46cf-aca3-93960c133428\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.690756 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6tsq\" (UniqueName: \"kubernetes.io/projected/1fb0006d-63be-4593-a64e-9af6225de71a-kube-api-access-b6tsq\") pod \"catalog-operator-68c6474976-w76w7\" (UID: \"1fb0006d-63be-4593-a64e-9af6225de71a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.691280 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm6dw\" (UniqueName: \"kubernetes.io/projected/20ccb5ed-eda9-459d-9581-ad5b6d02cb9d-kube-api-access-vm6dw\") pod \"olm-operator-6b444d44fb-vt4cz\" (UID: \"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.692282 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cd0711ac-4869-4b85-b204-941a95b922d3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5h6sf\" (UID: \"cd0711ac-4869-4b85-b204-941a95b922d3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.694816 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpw2t\" (UniqueName: \"kubernetes.io/projected/c0c1b3ff-a655-44aa-9060-af44630df2a3-kube-api-access-jpw2t\") pod \"machine-approver-56656f9798-x689f\" (UID: \"c0c1b3ff-a655-44aa-9060-af44630df2a3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.695364 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrkkd\" (UniqueName: \"kubernetes.io/projected/29764aad-227f-400b-9ebd-703c0e69f519-kube-api-access-nrkkd\") pod \"dns-operator-744455d44c-q5bbl\" (UID: \"29764aad-227f-400b-9ebd-703c0e69f519\") " pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.701702 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.715828 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722462 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5f2c2718-1e40-4ace-baa4-3d778124c0a3-webhook-cert\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722522 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j25f\" (UniqueName: \"kubernetes.io/projected/d347b264-47d4-4258-9473-ab7fb58a4cfa-kube-api-access-2j25f\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722536 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722564 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-trusted-ca\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722611 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722635 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/500b95ee-e25a-4aff-97f0-f8af3b1f8831-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722687 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81163ec6-106d-4159-b005-49b6fcdb1898-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w245c\" (UID: \"81163ec6-106d-4159-b005-49b6fcdb1898\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722718 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d347b264-47d4-4258-9473-ab7fb58a4cfa-config\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722737 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/500b95ee-e25a-4aff-97f0-f8af3b1f8831-audit-policies\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722779 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/500b95ee-e25a-4aff-97f0-f8af3b1f8831-encryption-config\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722815 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tqp4\" (UniqueName: \"kubernetes.io/projected/5f2c2718-1e40-4ace-baa4-3d778124c0a3-kube-api-access-7tqp4\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722839 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d347b264-47d4-4258-9473-ab7fb58a4cfa-etcd-service-ca\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722861 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81163ec6-106d-4159-b005-49b6fcdb1898-config\") pod \"kube-apiserver-operator-766d6c64bb-w245c\" (UID: \"81163ec6-106d-4159-b005-49b6fcdb1898\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722887 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d347b264-47d4-4258-9473-ab7fb58a4cfa-etcd-ca\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722926 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81163ec6-106d-4159-b005-49b6fcdb1898-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w245c\" (UID: \"81163ec6-106d-4159-b005-49b6fcdb1898\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.722953 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/500b95ee-e25a-4aff-97f0-f8af3b1f8831-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.723022 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-client-ca\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: E1210 09:36:24.723324 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:25.223308701 +0000 UTC m=+147.966855052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.723358 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-serving-cert\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.723593 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d347b264-47d4-4258-9473-ab7fb58a4cfa-serving-cert\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.723868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-config\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.723926 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/500b95ee-e25a-4aff-97f0-f8af3b1f8831-serving-cert\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.723964 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d347b264-47d4-4258-9473-ab7fb58a4cfa-etcd-client\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.723991 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/500b95ee-e25a-4aff-97f0-f8af3b1f8831-etcd-client\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724033 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhbx8\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-kube-api-access-zhbx8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724066 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ff8p\" (UniqueName: \"kubernetes.io/projected/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-kube-api-access-9ff8p\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724093 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/500b95ee-e25a-4aff-97f0-f8af3b1f8831-audit-dir\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724113 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-certificates\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724131 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bzp9\" (UniqueName: \"kubernetes.io/projected/d5e50315-240d-4034-885f-386587a9f062-kube-api-access-2bzp9\") pod \"service-ca-operator-777779d784-xshb8\" (UID: \"d5e50315-240d-4034-885f-386587a9f062\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724176 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724203 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-bound-sa-token\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724227 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-tls\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724243 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np8mx\" (UniqueName: \"kubernetes.io/projected/500b95ee-e25a-4aff-97f0-f8af3b1f8831-kube-api-access-np8mx\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724260 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5e50315-240d-4034-885f-386587a9f062-serving-cert\") pod \"service-ca-operator-777779d784-xshb8\" (UID: \"d5e50315-240d-4034-885f-386587a9f062\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724606 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724643 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5e50315-240d-4034-885f-386587a9f062-config\") pod \"service-ca-operator-777779d784-xshb8\" (UID: \"d5e50315-240d-4034-885f-386587a9f062\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724818 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5f2c2718-1e40-4ace-baa4-3d778124c0a3-apiservice-cert\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.724857 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5f2c2718-1e40-4ace-baa4-3d778124c0a3-tmpfs\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.733002 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825284 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825415 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-client-ca\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825440 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-serving-cert\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825490 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d347b264-47d4-4258-9473-ab7fb58a4cfa-serving-cert\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825517 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dnj8g\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825566 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0354cc8b-8f2b-4166-9a7c-c9ebeb59a979-certs\") pod \"machine-config-server-mm5xh\" (UID: \"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979\") " pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825588 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-config\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825607 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/500b95ee-e25a-4aff-97f0-f8af3b1f8831-serving-cert\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825626 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d347b264-47d4-4258-9473-ab7fb58a4cfa-etcd-client\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825643 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/500b95ee-e25a-4aff-97f0-f8af3b1f8831-etcd-client\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825663 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/851585ad-0aba-46c3-8161-b329fb65655b-proxy-tls\") pod \"machine-config-controller-84d6567774-25cwg\" (UID: \"851585ad-0aba-46c3-8161-b329fb65655b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825686 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhbx8\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-kube-api-access-zhbx8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825710 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tmj4\" (UniqueName: \"kubernetes.io/projected/851585ad-0aba-46c3-8161-b329fb65655b-kube-api-access-2tmj4\") pod \"machine-config-controller-84d6567774-25cwg\" (UID: \"851585ad-0aba-46c3-8161-b329fb65655b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825731 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ff8p\" (UniqueName: \"kubernetes.io/projected/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-kube-api-access-9ff8p\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825817 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/500b95ee-e25a-4aff-97f0-f8af3b1f8831-audit-dir\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825885 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-certificates\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825938 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bzp9\" (UniqueName: \"kubernetes.io/projected/d5e50315-240d-4034-885f-386587a9f062-kube-api-access-2bzp9\") pod \"service-ca-operator-777779d784-xshb8\" (UID: \"d5e50315-240d-4034-885f-386587a9f062\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.825989 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d1d191ca-a746-4352-8013-a9f51bd57f05-signing-key\") pod \"service-ca-9c57cc56f-nr6bd\" (UID: \"d1d191ca-a746-4352-8013-a9f51bd57f05\") " pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826043 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-csi-data-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826145 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826191 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/582cfff1-c3fb-462a-b80b-d14cd76a6c88-proxy-tls\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826264 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-bound-sa-token\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826290 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vgq2\" (UniqueName: \"kubernetes.io/projected/0354cc8b-8f2b-4166-9a7c-c9ebeb59a979-kube-api-access-7vgq2\") pod \"machine-config-server-mm5xh\" (UID: \"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979\") " pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826312 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q7mg\" (UniqueName: \"kubernetes.io/projected/f37a1d66-5d7d-4d58-9f23-401de6615848-kube-api-access-6q7mg\") pod \"marketplace-operator-79b997595-dnj8g\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826385 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-tls\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826434 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np8mx\" (UniqueName: \"kubernetes.io/projected/500b95ee-e25a-4aff-97f0-f8af3b1f8831-kube-api-access-np8mx\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5e50315-240d-4034-885f-386587a9f062-serving-cert\") pod \"service-ca-operator-777779d784-xshb8\" (UID: \"d5e50315-240d-4034-885f-386587a9f062\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826548 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826594 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5e50315-240d-4034-885f-386587a9f062-config\") pod \"service-ca-operator-777779d784-xshb8\" (UID: \"d5e50315-240d-4034-885f-386587a9f062\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826694 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwghf\" (UniqueName: \"kubernetes.io/projected/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-kube-api-access-gwghf\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826794 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5f2c2718-1e40-4ace-baa4-3d778124c0a3-apiservice-cert\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/582cfff1-c3fb-462a-b80b-d14cd76a6c88-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826893 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqp5g\" (UniqueName: \"kubernetes.io/projected/5cab43dc-7d86-4f09-8049-d6f8a8c67b75-kube-api-access-bqp5g\") pod \"ingress-canary-m8hpb\" (UID: \"5cab43dc-7d86-4f09-8049-d6f8a8c67b75\") " pod="openshift-ingress-canary/ingress-canary-m8hpb" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.826972 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dnj8g\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.827041 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5f2c2718-1e40-4ace-baa4-3d778124c0a3-tmpfs\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.827177 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5f2c2718-1e40-4ace-baa4-3d778124c0a3-webhook-cert\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.827230 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j25f\" (UniqueName: \"kubernetes.io/projected/d347b264-47d4-4258-9473-ab7fb58a4cfa-kube-api-access-2j25f\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.827228 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-client-ca\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: E1210 09:36:24.828536 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:25.328489191 +0000 UTC m=+148.072035442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.828578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-trusted-ca\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.828633 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/851585ad-0aba-46c3-8161-b329fb65655b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-25cwg\" (UID: \"851585ad-0aba-46c3-8161-b329fb65655b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.828701 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-registration-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.828787 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8c174d6b-eb6a-461c-a58b-7f497a5118e6-metrics-tls\") pod \"dns-default-hg2rw\" (UID: \"8c174d6b-eb6a-461c-a58b-7f497a5118e6\") " pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.828851 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d1d191ca-a746-4352-8013-a9f51bd57f05-signing-cabundle\") pod \"service-ca-9c57cc56f-nr6bd\" (UID: \"d1d191ca-a746-4352-8013-a9f51bd57f05\") " pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.828956 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.828990 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/500b95ee-e25a-4aff-97f0-f8af3b1f8831-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.829101 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81163ec6-106d-4159-b005-49b6fcdb1898-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w245c\" (UID: \"81163ec6-106d-4159-b005-49b6fcdb1898\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.829201 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d347b264-47d4-4258-9473-ab7fb58a4cfa-config\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.829233 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/500b95ee-e25a-4aff-97f0-f8af3b1f8831-audit-policies\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.829258 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nd27\" (UniqueName: \"kubernetes.io/projected/d1d191ca-a746-4352-8013-a9f51bd57f05-kube-api-access-9nd27\") pod \"service-ca-9c57cc56f-nr6bd\" (UID: \"d1d191ca-a746-4352-8013-a9f51bd57f05\") " pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.829292 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c174d6b-eb6a-461c-a58b-7f497a5118e6-config-volume\") pod \"dns-default-hg2rw\" (UID: \"8c174d6b-eb6a-461c-a58b-7f497a5118e6\") " pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.829314 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0354cc8b-8f2b-4166-9a7c-c9ebeb59a979-node-bootstrap-token\") pod \"machine-config-server-mm5xh\" (UID: \"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979\") " pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.829342 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/500b95ee-e25a-4aff-97f0-f8af3b1f8831-encryption-config\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.829365 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-plugins-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.829434 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tqp4\" (UniqueName: \"kubernetes.io/projected/5f2c2718-1e40-4ace-baa4-3d778124c0a3-kube-api-access-7tqp4\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.829460 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d347b264-47d4-4258-9473-ab7fb58a4cfa-etcd-service-ca\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.831402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-config\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.833040 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-certificates\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: E1210 09:36:24.837240 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:25.337221146 +0000 UTC m=+148.080767397 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.828967 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/500b95ee-e25a-4aff-97f0-f8af3b1f8831-audit-dir\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.838418 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/500b95ee-e25a-4aff-97f0-f8af3b1f8831-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.843489 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-serving-cert\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.846347 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-tls\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.922737 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5e50315-240d-4034-885f-386587a9f062-serving-cert\") pod \"service-ca-operator-777779d784-xshb8\" (UID: \"d5e50315-240d-4034-885f-386587a9f062\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.923640 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d347b264-47d4-4258-9473-ab7fb58a4cfa-serving-cert\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.924226 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/500b95ee-e25a-4aff-97f0-f8af3b1f8831-serving-cert\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.924549 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/500b95ee-e25a-4aff-97f0-f8af3b1f8831-encryption-config\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.932760 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d347b264-47d4-4258-9473-ab7fb58a4cfa-config\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.933356 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/500b95ee-e25a-4aff-97f0-f8af3b1f8831-audit-policies\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.934312 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.935171 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.942184 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.944686 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bzp9\" (UniqueName: \"kubernetes.io/projected/d5e50315-240d-4034-885f-386587a9f062-kube-api-access-2bzp9\") pod \"service-ca-operator-777779d784-xshb8\" (UID: \"d5e50315-240d-4034-885f-386587a9f062\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.945421 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5f2c2718-1e40-4ace-baa4-3d778124c0a3-tmpfs\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.945612 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.946217 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5e50315-240d-4034-885f-386587a9f062-config\") pod \"service-ca-operator-777779d784-xshb8\" (UID: \"d5e50315-240d-4034-885f-386587a9f062\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.946688 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.947399 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81163ec6-106d-4159-b005-49b6fcdb1898-config\") pod \"kube-apiserver-operator-766d6c64bb-w245c\" (UID: \"81163ec6-106d-4159-b005-49b6fcdb1898\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.947804 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d347b264-47d4-4258-9473-ab7fb58a4cfa-etcd-service-ca\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.948264 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.948725 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81163ec6-106d-4159-b005-49b6fcdb1898-config\") pod \"kube-apiserver-operator-766d6c64bb-w245c\" (UID: \"81163ec6-106d-4159-b005-49b6fcdb1898\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.948778 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cab43dc-7d86-4f09-8049-d6f8a8c67b75-cert\") pod \"ingress-canary-m8hpb\" (UID: \"5cab43dc-7d86-4f09-8049-d6f8a8c67b75\") " pod="openshift-ingress-canary/ingress-canary-m8hpb" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.948878 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.949008 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/582cfff1-c3fb-462a-b80b-d14cd76a6c88-images\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.949048 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-mountpoint-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.949888 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d347b264-47d4-4258-9473-ab7fb58a4cfa-etcd-ca\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.949953 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w99nm\" (UniqueName: \"kubernetes.io/projected/582cfff1-c3fb-462a-b80b-d14cd76a6c88-kube-api-access-w99nm\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.950276 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81163ec6-106d-4159-b005-49b6fcdb1898-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w245c\" (UID: \"81163ec6-106d-4159-b005-49b6fcdb1898\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.950763 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5f2c2718-1e40-4ace-baa4-3d778124c0a3-webhook-cert\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.950759 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-socket-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.951059 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sdm5\" (UniqueName: \"kubernetes.io/projected/8c174d6b-eb6a-461c-a58b-7f497a5118e6-kube-api-access-5sdm5\") pod \"dns-default-hg2rw\" (UID: \"8c174d6b-eb6a-461c-a58b-7f497a5118e6\") " pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.951122 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/500b95ee-e25a-4aff-97f0-f8af3b1f8831-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.951112 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d347b264-47d4-4258-9473-ab7fb58a4cfa-etcd-client\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.951775 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/500b95ee-e25a-4aff-97f0-f8af3b1f8831-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.952343 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d347b264-47d4-4258-9473-ab7fb58a4cfa-etcd-ca\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.952717 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5f2c2718-1e40-4ace-baa4-3d778124c0a3-apiservice-cert\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.953080 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-trusted-ca\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.956969 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81163ec6-106d-4159-b005-49b6fcdb1898-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w245c\" (UID: \"81163ec6-106d-4159-b005-49b6fcdb1898\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.957336 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/500b95ee-e25a-4aff-97f0-f8af3b1f8831-etcd-client\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.959220 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.982668 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j25f\" (UniqueName: \"kubernetes.io/projected/d347b264-47d4-4258-9473-ab7fb58a4cfa-kube-api-access-2j25f\") pod \"etcd-operator-b45778765-8wtv4\" (UID: \"d347b264-47d4-4258-9473-ab7fb58a4cfa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.985526 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np8mx\" (UniqueName: \"kubernetes.io/projected/500b95ee-e25a-4aff-97f0-f8af3b1f8831-kube-api-access-np8mx\") pod \"apiserver-7bbb656c7d-7g7c9\" (UID: \"500b95ee-e25a-4aff-97f0-f8af3b1f8831\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.987746 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" Dec 10 09:36:24 crc kubenswrapper[4715]: I1210 09:36:24.997872 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-bound-sa-token\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.013601 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81163ec6-106d-4159-b005-49b6fcdb1898-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w245c\" (UID: \"81163ec6-106d-4159-b005-49b6fcdb1898\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.044611 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052156 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052388 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-registration-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052426 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8c174d6b-eb6a-461c-a58b-7f497a5118e6-metrics-tls\") pod \"dns-default-hg2rw\" (UID: \"8c174d6b-eb6a-461c-a58b-7f497a5118e6\") " pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d1d191ca-a746-4352-8013-a9f51bd57f05-signing-cabundle\") pod \"service-ca-9c57cc56f-nr6bd\" (UID: \"d1d191ca-a746-4352-8013-a9f51bd57f05\") " pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052491 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nd27\" (UniqueName: \"kubernetes.io/projected/d1d191ca-a746-4352-8013-a9f51bd57f05-kube-api-access-9nd27\") pod \"service-ca-9c57cc56f-nr6bd\" (UID: \"d1d191ca-a746-4352-8013-a9f51bd57f05\") " pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052511 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c174d6b-eb6a-461c-a58b-7f497a5118e6-config-volume\") pod \"dns-default-hg2rw\" (UID: \"8c174d6b-eb6a-461c-a58b-7f497a5118e6\") " pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052567 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0354cc8b-8f2b-4166-9a7c-c9ebeb59a979-node-bootstrap-token\") pod \"machine-config-server-mm5xh\" (UID: \"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979\") " pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052589 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-plugins-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cab43dc-7d86-4f09-8049-d6f8a8c67b75-cert\") pod \"ingress-canary-m8hpb\" (UID: \"5cab43dc-7d86-4f09-8049-d6f8a8c67b75\") " pod="openshift-ingress-canary/ingress-canary-m8hpb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052681 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/582cfff1-c3fb-462a-b80b-d14cd76a6c88-images\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052736 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-mountpoint-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052756 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w99nm\" (UniqueName: \"kubernetes.io/projected/582cfff1-c3fb-462a-b80b-d14cd76a6c88-kube-api-access-w99nm\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052780 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-socket-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052800 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sdm5\" (UniqueName: \"kubernetes.io/projected/8c174d6b-eb6a-461c-a58b-7f497a5118e6-kube-api-access-5sdm5\") pod \"dns-default-hg2rw\" (UID: \"8c174d6b-eb6a-461c-a58b-7f497a5118e6\") " pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052822 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dnj8g\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052877 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0354cc8b-8f2b-4166-9a7c-c9ebeb59a979-certs\") pod \"machine-config-server-mm5xh\" (UID: \"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979\") " pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052903 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/851585ad-0aba-46c3-8161-b329fb65655b-proxy-tls\") pod \"machine-config-controller-84d6567774-25cwg\" (UID: \"851585ad-0aba-46c3-8161-b329fb65655b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052944 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tmj4\" (UniqueName: \"kubernetes.io/projected/851585ad-0aba-46c3-8161-b329fb65655b-kube-api-access-2tmj4\") pod \"machine-config-controller-84d6567774-25cwg\" (UID: \"851585ad-0aba-46c3-8161-b329fb65655b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052974 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d1d191ca-a746-4352-8013-a9f51bd57f05-signing-key\") pod \"service-ca-9c57cc56f-nr6bd\" (UID: \"d1d191ca-a746-4352-8013-a9f51bd57f05\") " pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.052997 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-csi-data-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.053052 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/582cfff1-c3fb-462a-b80b-d14cd76a6c88-proxy-tls\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.053075 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vgq2\" (UniqueName: \"kubernetes.io/projected/0354cc8b-8f2b-4166-9a7c-c9ebeb59a979-kube-api-access-7vgq2\") pod \"machine-config-server-mm5xh\" (UID: \"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979\") " pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.053100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q7mg\" (UniqueName: \"kubernetes.io/projected/f37a1d66-5d7d-4d58-9f23-401de6615848-kube-api-access-6q7mg\") pod \"marketplace-operator-79b997595-dnj8g\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.053129 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwghf\" (UniqueName: \"kubernetes.io/projected/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-kube-api-access-gwghf\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.053152 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/582cfff1-c3fb-462a-b80b-d14cd76a6c88-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.053222 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqp5g\" (UniqueName: \"kubernetes.io/projected/5cab43dc-7d86-4f09-8049-d6f8a8c67b75-kube-api-access-bqp5g\") pod \"ingress-canary-m8hpb\" (UID: \"5cab43dc-7d86-4f09-8049-d6f8a8c67b75\") " pod="openshift-ingress-canary/ingress-canary-m8hpb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.053257 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dnj8g\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.053307 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/851585ad-0aba-46c3-8161-b329fb65655b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-25cwg\" (UID: \"851585ad-0aba-46c3-8161-b329fb65655b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.053884 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d1d191ca-a746-4352-8013-a9f51bd57f05-signing-cabundle\") pod \"service-ca-9c57cc56f-nr6bd\" (UID: \"d1d191ca-a746-4352-8013-a9f51bd57f05\") " pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:25 crc kubenswrapper[4715]: E1210 09:36:25.054009 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:25.553991953 +0000 UTC m=+148.297538204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.054084 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-registration-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.054176 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c174d6b-eb6a-461c-a58b-7f497a5118e6-config-volume\") pod \"dns-default-hg2rw\" (UID: \"8c174d6b-eb6a-461c-a58b-7f497a5118e6\") " pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.054422 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/851585ad-0aba-46c3-8161-b329fb65655b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-25cwg\" (UID: \"851585ad-0aba-46c3-8161-b329fb65655b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.058126 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cab43dc-7d86-4f09-8049-d6f8a8c67b75-cert\") pod \"ingress-canary-m8hpb\" (UID: \"5cab43dc-7d86-4f09-8049-d6f8a8c67b75\") " pod="openshift-ingress-canary/ingress-canary-m8hpb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.058755 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0354cc8b-8f2b-4166-9a7c-c9ebeb59a979-node-bootstrap-token\") pod \"machine-config-server-mm5xh\" (UID: \"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979\") " pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.059011 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/582cfff1-c3fb-462a-b80b-d14cd76a6c88-images\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.059139 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-csi-data-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.062454 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/582cfff1-c3fb-462a-b80b-d14cd76a6c88-proxy-tls\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.063387 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/582cfff1-c3fb-462a-b80b-d14cd76a6c88-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.053216 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-plugins-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.068280 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-mountpoint-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.068439 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-socket-dir\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.079018 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0354cc8b-8f2b-4166-9a7c-c9ebeb59a979-certs\") pod \"machine-config-server-mm5xh\" (UID: \"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979\") " pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.080576 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8c174d6b-eb6a-461c-a58b-7f497a5118e6-metrics-tls\") pod \"dns-default-hg2rw\" (UID: \"8c174d6b-eb6a-461c-a58b-7f497a5118e6\") " pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.083466 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/851585ad-0aba-46c3-8161-b329fb65655b-proxy-tls\") pod \"machine-config-controller-84d6567774-25cwg\" (UID: \"851585ad-0aba-46c3-8161-b329fb65655b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.083947 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhbx8\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-kube-api-access-zhbx8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.086374 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nd27\" (UniqueName: \"kubernetes.io/projected/d1d191ca-a746-4352-8013-a9f51bd57f05-kube-api-access-9nd27\") pod \"service-ca-9c57cc56f-nr6bd\" (UID: \"d1d191ca-a746-4352-8013-a9f51bd57f05\") " pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.103925 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dnj8g\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.104053 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d1d191ca-a746-4352-8013-a9f51bd57f05-signing-key\") pod \"service-ca-9c57cc56f-nr6bd\" (UID: \"d1d191ca-a746-4352-8013-a9f51bd57f05\") " pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.104451 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tqp4\" (UniqueName: \"kubernetes.io/projected/5f2c2718-1e40-4ace-baa4-3d778124c0a3-kube-api-access-7tqp4\") pod \"packageserver-d55dfcdfc-nxdbh\" (UID: \"5f2c2718-1e40-4ace-baa4-3d778124c0a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.104574 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dnj8g\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.107822 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vgq2\" (UniqueName: \"kubernetes.io/projected/0354cc8b-8f2b-4166-9a7c-c9ebeb59a979-kube-api-access-7vgq2\") pod \"machine-config-server-mm5xh\" (UID: \"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979\") " pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.125770 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-mm5xh" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.134366 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ff8p\" (UniqueName: \"kubernetes.io/projected/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-kube-api-access-9ff8p\") pod \"route-controller-manager-6576b87f9c-z96cl\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.160585 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:25 crc kubenswrapper[4715]: E1210 09:36:25.161005 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:25.660988656 +0000 UTC m=+148.404534907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.207017 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqp5g\" (UniqueName: \"kubernetes.io/projected/5cab43dc-7d86-4f09-8049-d6f8a8c67b75-kube-api-access-bqp5g\") pod \"ingress-canary-m8hpb\" (UID: \"5cab43dc-7d86-4f09-8049-d6f8a8c67b75\") " pod="openshift-ingress-canary/ingress-canary-m8hpb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.209741 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sdm5\" (UniqueName: \"kubernetes.io/projected/8c174d6b-eb6a-461c-a58b-7f497a5118e6-kube-api-access-5sdm5\") pod \"dns-default-hg2rw\" (UID: \"8c174d6b-eb6a-461c-a58b-7f497a5118e6\") " pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.211546 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w99nm\" (UniqueName: \"kubernetes.io/projected/582cfff1-c3fb-462a-b80b-d14cd76a6c88-kube-api-access-w99nm\") pod \"machine-config-operator-74547568cd-h4crp\" (UID: \"582cfff1-c3fb-462a-b80b-d14cd76a6c88\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.215503 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwghf\" (UniqueName: \"kubernetes.io/projected/429ceb66-19e6-4fa6-bd59-dc06aa7c3d79-kube-api-access-gwghf\") pod \"csi-hostpathplugin-6njjb\" (UID: \"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79\") " pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.216221 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.216482 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q7mg\" (UniqueName: \"kubernetes.io/projected/f37a1d66-5d7d-4d58-9f23-401de6615848-kube-api-access-6q7mg\") pod \"marketplace-operator-79b997595-dnj8g\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.217685 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tmj4\" (UniqueName: \"kubernetes.io/projected/851585ad-0aba-46c3-8161-b329fb65655b-kube-api-access-2tmj4\") pod \"machine-config-controller-84d6567774-25cwg\" (UID: \"851585ad-0aba-46c3-8161-b329fb65655b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.222983 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.261845 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:25 crc kubenswrapper[4715]: E1210 09:36:25.262632 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:25.762607963 +0000 UTC m=+148.506154224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.266999 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.292359 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.356572 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.356987 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.363894 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:25 crc kubenswrapper[4715]: E1210 09:36:25.365175 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:25.865159256 +0000 UTC m=+148.608705507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.372890 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.405689 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.453101 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.453803 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6njjb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.454391 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-m8hpb" Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.467628 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:25 crc kubenswrapper[4715]: E1210 09:36:25.468049 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:25.968030779 +0000 UTC m=+148.711577030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.574893 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:25 crc kubenswrapper[4715]: E1210 09:36:25.575340 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:26.075325641 +0000 UTC m=+148.818871892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.676484 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:25 crc kubenswrapper[4715]: E1210 09:36:25.676755 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:26.176727311 +0000 UTC m=+148.920273562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.778953 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-72qbm" event={"ID":"fc7f7af1-1312-42c8-bc68-4bb42832b42d","Type":"ContainerStarted","Data":"96f7df0d4850673962d2c5444298118b7bde82ef8495ff8bbf6273db5d39856a"} Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.778995 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-72qbm" event={"ID":"fc7f7af1-1312-42c8-bc68-4bb42832b42d","Type":"ContainerStarted","Data":"bf79b9cbf20250e576add670da9dc487ee0f6235146a572f0c00153a669afbf8"} Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.779006 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-mm5xh" event={"ID":"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979","Type":"ContainerStarted","Data":"7e28b756af97e9beb9fd6ecf0f7134592bf4ced76e2787bf3ced9ac0885d9aaa"} Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.779015 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" event={"ID":"c0c1b3ff-a655-44aa-9060-af44630df2a3","Type":"ContainerStarted","Data":"cc7671ef2e6d338a0d376ab2045ebf0ba2b9226207a59abaef85e45decfac02c"} Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.779706 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:25 crc kubenswrapper[4715]: E1210 09:36:25.782203 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:26.282181019 +0000 UTC m=+149.025727270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.887875 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:25 crc kubenswrapper[4715]: E1210 09:36:25.890036 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:26.390013157 +0000 UTC m=+149.133559408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:25 crc kubenswrapper[4715]: I1210 09:36:25.992168 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:25 crc kubenswrapper[4715]: E1210 09:36:25.993099 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:26.493073475 +0000 UTC m=+149.236619726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.094589 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:26 crc kubenswrapper[4715]: E1210 09:36:26.095203 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:26.595162705 +0000 UTC m=+149.338709106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.196527 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:26 crc kubenswrapper[4715]: E1210 09:36:26.196850 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:26.696837773 +0000 UTC m=+149.440384024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.304831 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:26 crc kubenswrapper[4715]: E1210 09:36:26.305196 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:26.805178995 +0000 UTC m=+149.548725246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.354501 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-72qbm" podStartSLOduration=130.354487374 podStartE2EDuration="2m10.354487374s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:26.352332231 +0000 UTC m=+149.095878472" watchObservedRunningTime="2025-12-10 09:36:26.354487374 +0000 UTC m=+149.098033625" Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.406733 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:26 crc kubenswrapper[4715]: E1210 09:36:26.407103 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:26.90708726 +0000 UTC m=+149.650633511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.508499 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:26 crc kubenswrapper[4715]: E1210 09:36:26.508732 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.008706786 +0000 UTC m=+149.752253037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.611189 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:26 crc kubenswrapper[4715]: E1210 09:36:26.611591 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.111571529 +0000 UTC m=+149.855117780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.641148 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.712660 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:26 crc kubenswrapper[4715]: E1210 09:36:26.713004 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.212960748 +0000 UTC m=+149.956506999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.728381 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:26 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:26 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:26 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.728458 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.731393 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" event={"ID":"c0c1b3ff-a655-44aa-9060-af44630df2a3","Type":"ContainerStarted","Data":"6440d307f43049db04c4563fff83a51447dd9c671b24ea95b0c31dca7bd129a4"} Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.731445 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" event={"ID":"c0c1b3ff-a655-44aa-9060-af44630df2a3","Type":"ContainerStarted","Data":"15e098a0ac27bb65ba419810d0bed8b3f4288dc6c8bd2aeee1dad6b87f41d457"} Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.748641 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-mm5xh" event={"ID":"0354cc8b-8f2b-4166-9a7c-c9ebeb59a979","Type":"ContainerStarted","Data":"dfd2c1ee846fd429bfa8f7705a820d241f89a12c590f4d870a2530afff4be24e"} Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.763320 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x689f" podStartSLOduration=130.763288597 podStartE2EDuration="2m10.763288597s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:26.761643279 +0000 UTC m=+149.505189540" watchObservedRunningTime="2025-12-10 09:36:26.763288597 +0000 UTC m=+149.506834868" Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.782013 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-mm5xh" podStartSLOduration=4.781980283 podStartE2EDuration="4.781980283s" podCreationTimestamp="2025-12-10 09:36:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:26.776155433 +0000 UTC m=+149.519701704" watchObservedRunningTime="2025-12-10 09:36:26.781980283 +0000 UTC m=+149.525526554" Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.814202 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:26 crc kubenswrapper[4715]: E1210 09:36:26.821957 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.321938899 +0000 UTC m=+150.065485150 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.916644 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:26 crc kubenswrapper[4715]: E1210 09:36:26.917319 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.417281112 +0000 UTC m=+150.160827363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.970205 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fd9v5"] Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.976864 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-l7gt6"] Dec 10 09:36:26 crc kubenswrapper[4715]: I1210 09:36:26.983934 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5"] Dec 10 09:36:27 crc kubenswrapper[4715]: W1210 09:36:27.000060 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9ce5159_4cdb_47c7_9193_8adde2d4e2d0.slice/crio-7b06b220dbe3a3b04894c9825f920cb72e3f46bd191742ddd7b90066ce6e172e WatchSource:0}: Error finding container 7b06b220dbe3a3b04894c9825f920cb72e3f46bd191742ddd7b90066ce6e172e: Status 404 returned error can't find the container with id 7b06b220dbe3a3b04894c9825f920cb72e3f46bd191742ddd7b90066ce6e172e Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.018258 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.018714 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.518692242 +0000 UTC m=+150.262238493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.106409 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-p5sqh"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.119949 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.120028 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.62000454 +0000 UTC m=+150.363550791 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.120506 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.120845 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.620833174 +0000 UTC m=+150.364379425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.129763 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.133432 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7qnsk"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.135194 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.154549 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.157352 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zphsm"] Dec 10 09:36:27 crc kubenswrapper[4715]: W1210 09:36:27.165668 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7a77c53_aea3_40ab_8b5b_2c6c51b83ff1.slice/crio-632646a0ad77fc64faaa1e572dcb023922c9e5055e6be858d5e8d2ddab66efde WatchSource:0}: Error finding container 632646a0ad77fc64faaa1e572dcb023922c9e5055e6be858d5e8d2ddab66efde: Status 404 returned error can't find the container with id 632646a0ad77fc64faaa1e572dcb023922c9e5055e6be858d5e8d2ddab66efde Dec 10 09:36:27 crc kubenswrapper[4715]: W1210 09:36:27.194737 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf55a019d_460a_4b2e_b553_f8cd1166dd29.slice/crio-d4b013ed2c0bf1b994997344af7a7ff8ace2182d012deee31a0529abfb524956 WatchSource:0}: Error finding container d4b013ed2c0bf1b994997344af7a7ff8ace2182d012deee31a0529abfb524956: Status 404 returned error can't find the container with id d4b013ed2c0bf1b994997344af7a7ff8ace2182d012deee31a0529abfb524956 Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.221843 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.222197 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.722160661 +0000 UTC m=+150.465706912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.222340 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.222850 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.722829681 +0000 UTC m=+150.466375932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.324486 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.325445 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.825407125 +0000 UTC m=+150.568953496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.427121 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.427713 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:27.927686561 +0000 UTC m=+150.671233002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.529303 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.529953 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.029889244 +0000 UTC m=+150.773435515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.563414 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nr6bd"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.567897 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-q6645"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.579205 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.582924 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7lwft"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.587158 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q5bbl"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.587215 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.589883 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.593806 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.597227 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.599255 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.633483 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.634126 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.134098646 +0000 UTC m=+150.877644897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.648715 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:27 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:27 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:27 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.649109 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.684525 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.684561 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.684573 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.684586 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.684599 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-m8hpb"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.684610 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hg2rw"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.684624 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.694219 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.705704 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.711495 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnj8g"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.730209 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-j2l56"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.734980 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.735304 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.235287689 +0000 UTC m=+150.978833940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.737077 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.762047 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xshb8"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.767299 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" event={"ID":"136c04d9-6eec-4412-91c3-0c426e0b7746","Type":"ContainerStarted","Data":"8fbb738bf78502cca1344ed381ca0c5073247c90e504804f2c939634c1304578"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.773552 4715 generic.go:334] "Generic (PLEG): container finished" podID="53adb33e-d25b-412a-be71-82e3e6380a3e" containerID="722dea7826bafbb74ca2d4ced254dba4ef40db0f8aa1e6bbe6959316574c7f55" exitCode=0 Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.773637 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" event={"ID":"53adb33e-d25b-412a-be71-82e3e6380a3e","Type":"ContainerDied","Data":"722dea7826bafbb74ca2d4ced254dba4ef40db0f8aa1e6bbe6959316574c7f55"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.773674 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" event={"ID":"53adb33e-d25b-412a-be71-82e3e6380a3e","Type":"ContainerStarted","Data":"e5c6ba6ea4067e4ae25fc631a9535354bfaad2698e185b9afae3159c4ebad038"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.776393 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8wtv4"] Dec 10 09:36:27 crc kubenswrapper[4715]: W1210 09:36:27.777776 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod500b95ee_e25a_4aff_97f0_f8af3b1f8831.slice/crio-2e2df20fa0c1f44a7725299add0af1bd35140b5a907a071be9460a5177493706 WatchSource:0}: Error finding container 2e2df20fa0c1f44a7725299add0af1bd35140b5a907a071be9460a5177493706: Status 404 returned error can't find the container with id 2e2df20fa0c1f44a7725299add0af1bd35140b5a907a071be9460a5177493706 Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.789667 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9509a8cf600d59a5b9da16d4ca2b4b5833421be428a4da91b07d530918e5c2df"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.789714 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"649b0845eef6c64133180100c52182c2a420bf1d8ff20ed5734f47ba17c21544"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.794286 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" event={"ID":"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1","Type":"ContainerStarted","Data":"1554d8c21cf416b54f7d32cfe94d127aa45e890f2ca309621101f07a15b7bfe1"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.794345 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" event={"ID":"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1","Type":"ContainerStarted","Data":"632646a0ad77fc64faaa1e572dcb023922c9e5055e6be858d5e8d2ddab66efde"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.795331 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.799876 4715 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-7qnsk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.799943 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" podUID="e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.800172 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" event={"ID":"7d475c48-8e5c-48c3-9407-cce393aaca28","Type":"ContainerStarted","Data":"99b2b99a89fdae1bdf38666065f91e901212e64d1ec4d0420ea54a6e2878dfc1"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.800201 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" event={"ID":"7d475c48-8e5c-48c3-9407-cce393aaca28","Type":"ContainerStarted","Data":"c6b421477ff8bc1603099c8f297edef6d6ab6918f4c82f800a408c754ca30f79"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.807436 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" event={"ID":"851585ad-0aba-46c3-8161-b329fb65655b","Type":"ContainerStarted","Data":"20673f318925eae1c2276d7a6172815204dce5c22287c71422b8a6e432cc4c70"} Dec 10 09:36:27 crc kubenswrapper[4715]: W1210 09:36:27.807675 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podefb368e0_756b_4323_8ca2_0852ff385eb4.slice/crio-3c5652c18d9c52ccd46ead026ad93beb3bbf377d4cb79748b16b6b1ddf36d4ac WatchSource:0}: Error finding container 3c5652c18d9c52ccd46ead026ad93beb3bbf377d4cb79748b16b6b1ddf36d4ac: Status 404 returned error can't find the container with id 3c5652c18d9c52ccd46ead026ad93beb3bbf377d4cb79748b16b6b1ddf36d4ac Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.819579 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" event={"ID":"39235b98-6c0c-4285-85f3-1137bc4c6fef","Type":"ContainerStarted","Data":"1b2decbc0c9016a28f2f6b0a7693baa1e7deb341b22714e72fe59ede4cbccf67"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.820953 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" event={"ID":"39235b98-6c0c-4285-85f3-1137bc4c6fef","Type":"ContainerStarted","Data":"1cb6f0cae90f93d0dba030658d0a39889133643d5bfdfc21f38b3396ab20a11c"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.823010 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" event={"ID":"d1d191ca-a746-4352-8013-a9f51bd57f05","Type":"ContainerStarted","Data":"1636ff0d59e1442977df8a1ab5a183a70e89caa6e2d64bde2a0c11ef468ec18f"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.828106 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" event={"ID":"45bf2bc4-681f-46cf-aca3-93960c133428","Type":"ContainerStarted","Data":"4ffdc2208b1da332fca5bbaebc842cec87e0fdba389eae87a7f1a25d02f3e5ef"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.832849 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" event={"ID":"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0","Type":"ContainerStarted","Data":"ed937bb5e0361f4e9ff4a1c3dc51a83a0179f571495f056406a752456707388d"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.832901 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" event={"ID":"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0","Type":"ContainerStarted","Data":"7b06b220dbe3a3b04894c9825f920cb72e3f46bd191742ddd7b90066ce6e172e"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.835891 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.845565 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.345539167 +0000 UTC m=+151.089085418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.854744 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" event={"ID":"b06279e0-9b31-40fb-ac5b-d7fd85abd9c8","Type":"ContainerStarted","Data":"3ea96059c817ebfea43e30a7d2d795af9e47ba39c5ff9283518dbac32bf82e1d"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.881478 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.888928 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" event={"ID":"f55a019d-460a-4b2e-b553-f8cd1166dd29","Type":"ContainerStarted","Data":"33962fe8dc377f8ed018e4c74c70b6f5e91eefb090e67c4c1305d8fe75e8c1ec"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.891841 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" event={"ID":"f55a019d-460a-4b2e-b553-f8cd1166dd29","Type":"ContainerStarted","Data":"d4b013ed2c0bf1b994997344af7a7ff8ace2182d012deee31a0529abfb524956"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.901253 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" event={"ID":"cd0711ac-4869-4b85-b204-941a95b922d3","Type":"ContainerStarted","Data":"1ff63ba1f0772cdd73c45ecb867a41512e5c081e2f644fead691d391e57b819c"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.905045 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.906386 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" event={"ID":"eff08ce1-5a34-48f6-9f72-1e548f563b20","Type":"ContainerStarted","Data":"6d2cfb45fc605138a23086541ff36c0a433e941582f5b67805ac0113ccc99b6d"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.907138 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6njjb"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.909413 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" event={"ID":"352662c6-4463-4850-89f1-e3aa2fbead2f","Type":"ContainerStarted","Data":"3ea0b5912db54321a738517bd49b686d8c7d50c850504936dd48ad2840e34331"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.910544 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" event={"ID":"81163ec6-106d-4159-b005-49b6fcdb1898","Type":"ContainerStarted","Data":"2a92ba8423e3838eb4020b7a879e526e074d32b92b6b8149a6a7b9e4657627da"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.923570 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" event={"ID":"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2","Type":"ContainerStarted","Data":"5218c4f490680fe96494b692ba0c35823a07ba407a69d9b2fc95472f797e9589"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.925192 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.931531 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" event={"ID":"29764aad-227f-400b-9ebd-703c0e69f519","Type":"ContainerStarted","Data":"2c5956450d7dac91b14ddccc414e8734aaa678b2936e60a483073c912cb2da16"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.938592 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.938759 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.438737137 +0000 UTC m=+151.182283388 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.938856 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:27 crc kubenswrapper[4715]: E1210 09:36:27.945547 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.445493704 +0000 UTC m=+151.189039955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.949391 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" event={"ID":"a1116225-4343-4dcf-b128-83a355de274e","Type":"ContainerStarted","Data":"467632bde6a0c81b22d126d0422841e0cfce38636df158acf94d4cf16fec400b"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.958231 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2a88f9afd025cd2c9e82b7d1384dc6ae43a2fcc2d765d959bb319b0613e9aba1"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.958292 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"4800e8cbc508164788139e12503a5f12c5ebda65afabe6ab8e16b9981ce93d76"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.962783 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b9k84"] Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.977052 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" event={"ID":"8a456b48-1823-4211-b1f9-cf64a5645e6a","Type":"ContainerStarted","Data":"dd8215f0fd02f37fb0d0468e413463cc579b7a3e1d964a3f44ac729c1698cbd9"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.993365 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zphsm" event={"ID":"5fbc3706-b2a3-48aa-a950-dc403907fa02","Type":"ContainerStarted","Data":"9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14"} Dec 10 09:36:27 crc kubenswrapper[4715]: I1210 09:36:27.993435 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zphsm" event={"ID":"5fbc3706-b2a3-48aa-a950-dc403907fa02","Type":"ContainerStarted","Data":"053328968d72f359580dbec95909ef7d7ae9e5580ccf916ec1c5ec658a9e9f07"} Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.006013 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3db1564dcb13c9bf2bcfe1b7f3cbca1b01b1c6809327bc928f02afb1ee3bb15a"} Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.006079 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"98250d71ee7918cf9a69f52dd415b4ccb0ec173254bef178a944050b02e1a74f"} Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.006732 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.011423 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7lwft" event={"ID":"be3075d3-3e88-4e8c-9aa7-41d7d5345290","Type":"ContainerStarted","Data":"7d23a70d538883fd2cf98fec15fa4d89034bcce6af09d4b460584ac43cca913b"} Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.015647 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5" event={"ID":"4435aa00-3653-4733-90ef-86f45701c917","Type":"ContainerStarted","Data":"d293127a6daf4422c73e13a7cf9fe48f778367293f92a4d9aebb0fbbdd17dbce"} Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.018436 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" event={"ID":"834668e1-7a44-4844-b2f6-3b629ceeba22","Type":"ContainerStarted","Data":"eaf82e6dffeca80f209cd8107adadcfe2efad8ba1fe9653629e32fecc9ae2d08"} Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.018486 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" event={"ID":"834668e1-7a44-4844-b2f6-3b629ceeba22","Type":"ContainerStarted","Data":"721fb03eee0a5b5f7202bb91c2db48bb69581198d99f49704a4da75500552316"} Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.021346 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" event={"ID":"f37a1d66-5d7d-4d58-9f23-401de6615848","Type":"ContainerStarted","Data":"48b484192210a42345e77cd3119ac79f0772d799c7d0f250694d47912cdf42b1"} Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.043037 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:28 crc kubenswrapper[4715]: E1210 09:36:28.043953 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.543928138 +0000 UTC m=+151.287474389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.146138 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:28 crc kubenswrapper[4715]: E1210 09:36:28.147247 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.647218863 +0000 UTC m=+151.390765114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.194641 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-zphsm" podStartSLOduration=132.194612426 podStartE2EDuration="2m12.194612426s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:28.191520846 +0000 UTC m=+150.935067097" watchObservedRunningTime="2025-12-10 09:36:28.194612426 +0000 UTC m=+150.938158677" Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.219961 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-gvqxr" podStartSLOduration=132.219937165 podStartE2EDuration="2m12.219937165s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:28.216961528 +0000 UTC m=+150.960507779" watchObservedRunningTime="2025-12-10 09:36:28.219937165 +0000 UTC m=+150.963483416" Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.257764 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:28 crc kubenswrapper[4715]: E1210 09:36:28.258259 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.758233623 +0000 UTC m=+151.501779874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.359477 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:28 crc kubenswrapper[4715]: E1210 09:36:28.359958 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.859941792 +0000 UTC m=+151.603488043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.442090 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-l7gt6" podStartSLOduration=132.442076739 podStartE2EDuration="2m12.442076739s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:28.441129831 +0000 UTC m=+151.184676082" watchObservedRunningTime="2025-12-10 09:36:28.442076739 +0000 UTC m=+151.185622990" Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.460903 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:28 crc kubenswrapper[4715]: E1210 09:36:28.463037 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:28.96301515 +0000 UTC m=+151.706561401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.527888 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" podStartSLOduration=132.527870184 podStartE2EDuration="2m12.527870184s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:28.50649922 +0000 UTC m=+151.250045491" watchObservedRunningTime="2025-12-10 09:36:28.527870184 +0000 UTC m=+151.271416435" Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.546925 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w2jc5" podStartSLOduration=132.546891709 podStartE2EDuration="2m12.546891709s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:28.526859554 +0000 UTC m=+151.270405795" watchObservedRunningTime="2025-12-10 09:36:28.546891709 +0000 UTC m=+151.290437960" Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.578115 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:28 crc kubenswrapper[4715]: E1210 09:36:28.578372 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.078362327 +0000 UTC m=+151.821908578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.652493 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:28 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:28 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:28 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.653027 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.680585 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:28 crc kubenswrapper[4715]: E1210 09:36:28.681371 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.181346433 +0000 UTC m=+151.924892684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.783298 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:28 crc kubenswrapper[4715]: E1210 09:36:28.783809 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.283791924 +0000 UTC m=+152.027338175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:28 crc kubenswrapper[4715]: I1210 09:36:28.912964 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:28 crc kubenswrapper[4715]: E1210 09:36:28.913392 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.413371786 +0000 UTC m=+152.156918037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.014623 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.015052 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.515035544 +0000 UTC m=+152.258581795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.039764 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" event={"ID":"412a7188-42b3-4f08-affa-f4df0c076e46","Type":"ContainerStarted","Data":"ec0cec9516381a2e5dd90ce6ec8bcd82b4a471bf2fcf3f1bd7aa092bb902cb7d"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.041706 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" event={"ID":"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d","Type":"ContainerStarted","Data":"1f93c26776658f1ecbb4bde6545abf2e3bda5f731ade74797b35c8de2e93c158"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.042282 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" event={"ID":"20ccb5ed-eda9-459d-9581-ad5b6d02cb9d","Type":"ContainerStarted","Data":"b1f8ab07780f154f2fda0bd9e78ab4d16d2d676849aca824d58e71071d341a3a"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.042595 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.053200 4715 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-vt4cz container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.053532 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" podUID="20ccb5ed-eda9-459d-9581-ad5b6d02cb9d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.055853 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" event={"ID":"5f2c2718-1e40-4ace-baa4-3d778124c0a3","Type":"ContainerStarted","Data":"bc8bb531bb77246bf2c751ac99889f8d7f961f7bbba8f1533674148a61bd4260"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.061546 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" podStartSLOduration=132.061521081 podStartE2EDuration="2m12.061521081s" podCreationTimestamp="2025-12-10 09:34:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:29.059626635 +0000 UTC m=+151.803172886" watchObservedRunningTime="2025-12-10 09:36:29.061521081 +0000 UTC m=+151.805067332" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.075138 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" event={"ID":"d347b264-47d4-4258-9473-ab7fb58a4cfa","Type":"ContainerStarted","Data":"8796cb722384d645a1e279e32e0a99ad93ff0356cff84ea7e690e95b69bfebba"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.077199 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hg2rw" event={"ID":"8c174d6b-eb6a-461c-a58b-7f497a5118e6","Type":"ContainerStarted","Data":"a1d54d16f6b048197b01939f9dd8a6235be20e323acdcbe9b578a1288be380c8"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.079509 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" event={"ID":"a1116225-4343-4dcf-b128-83a355de274e","Type":"ContainerStarted","Data":"17c8f08137560656c03e28b5ec3ebb0138888a5d46e5a2fcb4993211898b458a"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.103524 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nw2b" podStartSLOduration=133.103504586 podStartE2EDuration="2m13.103504586s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:29.102474916 +0000 UTC m=+151.846021177" watchObservedRunningTime="2025-12-10 09:36:29.103504586 +0000 UTC m=+151.847050847" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.103929 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" event={"ID":"d1d191ca-a746-4352-8013-a9f51bd57f05","Type":"ContainerStarted","Data":"42d2090dd50c30e4b355ab319965b437da34486a7883e972084c6e0d1eef579f"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.119658 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.119788 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.619767671 +0000 UTC m=+152.363313922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.120008 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.120337 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.620326407 +0000 UTC m=+152.363872658 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.124686 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" event={"ID":"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2","Type":"ContainerStarted","Data":"2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.125205 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.126601 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-m8hpb" event={"ID":"5cab43dc-7d86-4f09-8049-d6f8a8c67b75","Type":"ContainerStarted","Data":"03f1c263d25023bee80bd2d3dfbfa42be89b71531ac0cdfb1d724d23e6f36217"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.128150 4715 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-z96cl container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.128197 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" podUID="526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.141196 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-nr6bd" podStartSLOduration=132.141182366 podStartE2EDuration="2m12.141182366s" podCreationTimestamp="2025-12-10 09:34:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:29.13960504 +0000 UTC m=+151.883151291" watchObservedRunningTime="2025-12-10 09:36:29.141182366 +0000 UTC m=+151.884728617" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.152635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" event={"ID":"c9ce5159-4cdb-47c7-9193-8adde2d4e2d0","Type":"ContainerStarted","Data":"167aa80d68e5a5d91b8bbdb6c2e20e3015a2d1d462fe6b6c87a857bb73f70442"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.157195 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" event={"ID":"1fb0006d-63be-4593-a64e-9af6225de71a","Type":"ContainerStarted","Data":"af517e801d255153a02198d1daf47827ac17e010b0491c99b3db63bf27fea66f"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.158726 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" event={"ID":"851585ad-0aba-46c3-8161-b329fb65655b","Type":"ContainerStarted","Data":"5d36999b48c4c265b0aaec01aed8434495428de1ee5eff1aea8fc6e4ef4072ca"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.168387 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" event={"ID":"52a7659e-76f5-49b6-abf8-c3facab139ab","Type":"ContainerStarted","Data":"6dd7f7e833628245bea66328a295d6b32e8c6ee612e1114052382746abe7469b"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.169535 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" event={"ID":"8a456b48-1823-4211-b1f9-cf64a5645e6a","Type":"ContainerStarted","Data":"20cb3bd0a4f34f2eec73646f844ff76e8d7ab64fb8e3d3acf48a3bbe569e48b1"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.170188 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" event={"ID":"500b95ee-e25a-4aff-97f0-f8af3b1f8831","Type":"ContainerStarted","Data":"2e2df20fa0c1f44a7725299add0af1bd35140b5a907a071be9460a5177493706"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.170794 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6njjb" event={"ID":"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79","Type":"ContainerStarted","Data":"4e82fd7570eb4fb8b4b5e091214deccbf6afacc88e4b8dca1f5bb1e84efaf763"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.171630 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5" event={"ID":"4435aa00-3653-4733-90ef-86f45701c917","Type":"ContainerStarted","Data":"f53bd1f3ef21f9acd8f8a77986c660e7be68db02fc811aaa9f219add686eff3d"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.178731 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" podStartSLOduration=132.178712721 podStartE2EDuration="2m12.178712721s" podCreationTimestamp="2025-12-10 09:34:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:29.176704123 +0000 UTC m=+151.920250394" watchObservedRunningTime="2025-12-10 09:36:29.178712721 +0000 UTC m=+151.922258972" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.179233 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7lwft" event={"ID":"be3075d3-3e88-4e8c-9aa7-41d7d5345290","Type":"ContainerStarted","Data":"eb2f27cc04e7f737a8b5a68e8af59013e5dd38d55225fd4d21d9d985b92e47bf"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.179503 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.181530 4715 patch_prober.go:28] interesting pod/console-operator-58897d9998-7lwft container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.181559 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7lwft" podUID="be3075d3-3e88-4e8c-9aa7-41d7d5345290" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.182414 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b9k84" event={"ID":"55cadec6-a68b-4d03-abc0-d9c628c3e4d0","Type":"ContainerStarted","Data":"f3cbcb34f79a6db38e4405a3ad309e378823e9705a241f407fa621d47a32cbcf"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.210666 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-fd9v5" podStartSLOduration=133.210620683 podStartE2EDuration="2m13.210620683s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:29.207024368 +0000 UTC m=+151.950570649" watchObservedRunningTime="2025-12-10 09:36:29.210620683 +0000 UTC m=+151.954166954" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.215375 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" event={"ID":"45bf2bc4-681f-46cf-aca3-93960c133428","Type":"ContainerStarted","Data":"6bd457778fe023dc49a6332f4ba4f9ebd9291b2dd655bd61a3f1e70d02d37835"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.222011 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.222175 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.722147059 +0000 UTC m=+152.465693320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.222507 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.225266 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.72524781 +0000 UTC m=+152.468794061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.227831 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" event={"ID":"136c04d9-6eec-4412-91c3-0c426e0b7746","Type":"ContainerDied","Data":"461a5f64445ce5caf57014debf1052b1a4a9218567e9d1d14e4ef9c79e3f7fb1"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.227938 4715 generic.go:334] "Generic (PLEG): container finished" podID="136c04d9-6eec-4412-91c3-0c426e0b7746" containerID="461a5f64445ce5caf57014debf1052b1a4a9218567e9d1d14e4ef9c79e3f7fb1" exitCode=0 Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.253300 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" event={"ID":"582cfff1-c3fb-462a-b80b-d14cd76a6c88","Type":"ContainerStarted","Data":"5dea661ef8caa969dc4f2947ef7fd5e664c738b1ddcd16fbe6f9437a6d24a419"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.264885 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" event={"ID":"d5e50315-240d-4034-885f-386587a9f062","Type":"ContainerStarted","Data":"c10e729f4ec958b4d1b7146ae4aca703902104c9bbab66dab5a7cf78775c200d"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.272511 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tnrn5" podStartSLOduration=133.272486269 podStartE2EDuration="2m13.272486269s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:29.261641292 +0000 UTC m=+152.005187533" watchObservedRunningTime="2025-12-10 09:36:29.272486269 +0000 UTC m=+152.016032520" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.273146 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-7lwft" podStartSLOduration=133.273141558 podStartE2EDuration="2m13.273141558s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:29.226600679 +0000 UTC m=+151.970146930" watchObservedRunningTime="2025-12-10 09:36:29.273141558 +0000 UTC m=+152.016687809" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.317369 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" event={"ID":"eff08ce1-5a34-48f6-9f72-1e548f563b20","Type":"ContainerStarted","Data":"b5aad762bd37a1e7dc07b3550cd4a1bc33d28a93d1f23cffb1a73b6f40403908"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.317839 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.332869 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.333167 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.833130929 +0000 UTC m=+152.576677190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.333536 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.336029 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.836008963 +0000 UTC m=+152.579555214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.357988 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" podStartSLOduration=133.357972184 podStartE2EDuration="2m13.357972184s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:29.355032018 +0000 UTC m=+152.098578269" watchObservedRunningTime="2025-12-10 09:36:29.357972184 +0000 UTC m=+152.101518435" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.374473 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" event={"ID":"efb368e0-756b-4323-8ca2-0852ff385eb4","Type":"ContainerStarted","Data":"3c5652c18d9c52ccd46ead026ad93beb3bbf377d4cb79748b16b6b1ddf36d4ac"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.384885 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" event={"ID":"352662c6-4463-4850-89f1-e3aa2fbead2f","Type":"ContainerStarted","Data":"a1ccb24639f916c8ae53f28e0c73667e1e4b32d1477bd196c4a4939161c8ac21"} Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.403002 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.495041 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.495203 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.995176569 +0000 UTC m=+152.738722820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.495417 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.498213 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:29.998203377 +0000 UTC m=+152.741749628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.542255 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.597844 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.598076 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.098036501 +0000 UTC m=+152.841582762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.598678 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.600681 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.100659108 +0000 UTC m=+152.844205359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.646411 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:29 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:29 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:29 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.646880 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.709652 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.710092 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.210069381 +0000 UTC m=+152.953615632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.813615 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.813956 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.313941553 +0000 UTC m=+153.057487804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:29 crc kubenswrapper[4715]: I1210 09:36:29.914988 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:29 crc kubenswrapper[4715]: E1210 09:36:29.915612 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.41558855 +0000 UTC m=+153.159134801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.018663 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.019524 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.519496834 +0000 UTC m=+153.263043085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.120641 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.121028 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.620987666 +0000 UTC m=+153.364533907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.121164 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.121532 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.621503081 +0000 UTC m=+153.365049342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.221670 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.223037 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.723013324 +0000 UTC m=+153.466559595 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.324037 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.324421 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.824407744 +0000 UTC m=+153.567953995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.423001 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hg2rw" event={"ID":"8c174d6b-eb6a-461c-a58b-7f497a5118e6","Type":"ContainerStarted","Data":"b7b5a40ea0cf65ae3f3cd0d6d579328d8f80041b6a16d79d9325cf9fc4878813"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.425071 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.425458 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:30.925436253 +0000 UTC m=+153.668982504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.485795 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" event={"ID":"39235b98-6c0c-4285-85f3-1137bc4c6fef","Type":"ContainerStarted","Data":"89ee6b9bf61198d8265081e7b24f637e8847c58a080dab1b8a182c02cd707f0b"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.486744 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.503274 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" event={"ID":"efb368e0-756b-4323-8ca2-0852ff385eb4","Type":"ContainerStarted","Data":"ba038ccc002ae5e69cf5603e7af827400c078a0a8a5db976d6d1797f0bbdaa48"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.528838 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.529271 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:31.029256744 +0000 UTC m=+153.772802995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.533494 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" event={"ID":"f37a1d66-5d7d-4d58-9f23-401de6615848","Type":"ContainerStarted","Data":"dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.534511 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.543491 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dnj8g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.543602 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" podUID="f37a1d66-5d7d-4d58-9f23-401de6615848" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.557876 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" podStartSLOduration=133.557855318 podStartE2EDuration="2m13.557855318s" podCreationTimestamp="2025-12-10 09:34:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:30.527612246 +0000 UTC m=+153.271158497" watchObservedRunningTime="2025-12-10 09:36:30.557855318 +0000 UTC m=+153.301401569" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.561360 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" event={"ID":"5f2c2718-1e40-4ace-baa4-3d778124c0a3","Type":"ContainerStarted","Data":"faff7c05b6eddea811b514b73f83baa9c131b23aa0c20ee9faa031dfe2014da5"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.564207 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.569253 4715 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nxdbh container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" start-of-body= Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.569619 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" podUID="5f2c2718-1e40-4ace-baa4-3d778124c0a3" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.596425 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" event={"ID":"b06279e0-9b31-40fb-ac5b-d7fd85abd9c8","Type":"ContainerStarted","Data":"c65321dc00516fec035c91093495088f58828b6ae9f8b6b19809ebff9c98ce7c"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.597022 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" podStartSLOduration=134.597010061 podStartE2EDuration="2m14.597010061s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:30.55927648 +0000 UTC m=+153.302822731" watchObservedRunningTime="2025-12-10 09:36:30.597010061 +0000 UTC m=+153.340556312" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.614117 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" event={"ID":"d5e50315-240d-4034-885f-386587a9f062","Type":"ContainerStarted","Data":"6f720603b84a653c76dfadaf7fc6799540511b47605702ce21cb337c7a333883"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.632009 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" podStartSLOduration=134.631991032 podStartE2EDuration="2m14.631991032s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:30.60038646 +0000 UTC m=+153.343932711" watchObservedRunningTime="2025-12-10 09:36:30.631991032 +0000 UTC m=+153.375537283" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.634579 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.635728 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:31.135702461 +0000 UTC m=+153.879248702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.640200 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-m8hpb" event={"ID":"5cab43dc-7d86-4f09-8049-d6f8a8c67b75","Type":"ContainerStarted","Data":"3bcbebadc219c6a00ef95268f265f96b8bf17b85bec6b31857eef2ef1a0e7252"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.651644 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:30 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:30 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:30 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.651696 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.660965 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" event={"ID":"81163ec6-106d-4159-b005-49b6fcdb1898","Type":"ContainerStarted","Data":"ba869c8cb9d16a81aa613a6f329597bec842ef7e718b56be57804321f3a9f5aa"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.681113 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" event={"ID":"53adb33e-d25b-412a-be71-82e3e6380a3e","Type":"ContainerStarted","Data":"09f7bf7896466b7a9e615921de072ebbca61ca887cd42feea763bc40f88400bc"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.698093 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" podStartSLOduration=133.698074181 podStartE2EDuration="2m13.698074181s" podCreationTimestamp="2025-12-10 09:34:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:30.651468391 +0000 UTC m=+153.395014642" watchObservedRunningTime="2025-12-10 09:36:30.698074181 +0000 UTC m=+153.441620432" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.705408 4715 generic.go:334] "Generic (PLEG): container finished" podID="500b95ee-e25a-4aff-97f0-f8af3b1f8831" containerID="6df3eee736974597c80f84e35d70b43a729310227318e72e9ece978bca8a6f8b" exitCode=0 Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.706250 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" event={"ID":"500b95ee-e25a-4aff-97f0-f8af3b1f8831","Type":"ContainerDied","Data":"6df3eee736974597c80f84e35d70b43a729310227318e72e9ece978bca8a6f8b"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.736385 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.738834 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" event={"ID":"29764aad-227f-400b-9ebd-703c0e69f519","Type":"ContainerStarted","Data":"d68ee4a2ee348907f04d9bc139825ad2e0a4c6cb8dde89c792bceee8a57913c6"} Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.740291 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:31.240263873 +0000 UTC m=+153.983810184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.761974 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshb8" podStartSLOduration=133.761959976 podStartE2EDuration="2m13.761959976s" podCreationTimestamp="2025-12-10 09:34:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:30.699611756 +0000 UTC m=+153.443157997" watchObservedRunningTime="2025-12-10 09:36:30.761959976 +0000 UTC m=+153.505506227" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.763234 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-m8hpb" podStartSLOduration=8.763227913 podStartE2EDuration="8.763227913s" podCreationTimestamp="2025-12-10 09:36:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:30.745469035 +0000 UTC m=+153.489015296" watchObservedRunningTime="2025-12-10 09:36:30.763227913 +0000 UTC m=+153.506774164" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.767123 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" event={"ID":"cd0711ac-4869-4b85-b204-941a95b922d3","Type":"ContainerStarted","Data":"a0e319f2a2980e9bb48364a9f2d63442bfad8c8012b91ca1626bc001c7e5bd6b"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.793708 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b9k84" event={"ID":"55cadec6-a68b-4d03-abc0-d9c628c3e4d0","Type":"ContainerStarted","Data":"77ec3149beda57df0ce8099444cc73647c2202b18cf33a8722bf5c09999040df"} Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.794074 4715 patch_prober.go:28] interesting pod/console-operator-58897d9998-7lwft container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.794120 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7lwft" podUID="be3075d3-3e88-4e8c-9aa7-41d7d5345290" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.795004 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-b9k84" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.807876 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.808119 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.817554 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vt4cz" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.837104 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.838078 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:31.338046437 +0000 UTC m=+154.081592698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.842191 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w245c" podStartSLOduration=134.842176528 podStartE2EDuration="2m14.842176528s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:30.793041203 +0000 UTC m=+153.536587474" watchObservedRunningTime="2025-12-10 09:36:30.842176528 +0000 UTC m=+153.585722779" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.873322 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5h6sf" podStartSLOduration=134.873283686 podStartE2EDuration="2m14.873283686s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:30.873254955 +0000 UTC m=+153.616801206" watchObservedRunningTime="2025-12-10 09:36:30.873283686 +0000 UTC m=+153.616829937" Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.941271 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:30 crc kubenswrapper[4715]: E1210 09:36:30.945773 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:31.445756571 +0000 UTC m=+154.189302822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:30 crc kubenswrapper[4715]: I1210 09:36:30.951134 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-b9k84" podStartSLOduration=134.951111737 podStartE2EDuration="2m14.951111737s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:30.920319889 +0000 UTC m=+153.663866160" watchObservedRunningTime="2025-12-10 09:36:30.951111737 +0000 UTC m=+153.694657988" Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.043386 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.043773 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:31.543752682 +0000 UTC m=+154.287298933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.148167 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.148602 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:31.648584051 +0000 UTC m=+154.392130302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.233246 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.250511 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.250809 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:31.750790155 +0000 UTC m=+154.494336406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.353522 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.354357 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:31.854335027 +0000 UTC m=+154.597881278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.460471 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.460972 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:31.960946628 +0000 UTC m=+154.704492879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.562654 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.563028 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.063015348 +0000 UTC m=+154.806561599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.644256 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:31 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:31 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:31 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.644309 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.664163 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.664440 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.164407097 +0000 UTC m=+154.907953358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.664574 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.664973 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.164955363 +0000 UTC m=+154.908501624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.765520 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.765713 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.265680633 +0000 UTC m=+155.009226884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.765769 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.766202 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.266186638 +0000 UTC m=+155.009732879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.867305 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.867668 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.36764376 +0000 UTC m=+155.111190021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.868007 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.868371 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.368361111 +0000 UTC m=+155.111907362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.914479 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" event={"ID":"52a7659e-76f5-49b6-abf8-c3facab139ab","Type":"ContainerStarted","Data":"47a23117699d88a49d09593b5c53194756ec34bf3eda5cae90712d091f3334d2"} Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.933654 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" event={"ID":"352662c6-4463-4850-89f1-e3aa2fbead2f","Type":"ContainerStarted","Data":"3efd1c70334683574434fef36e257e747c5b3ceee3c1e4ef1ffb2a55e231f7d6"} Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.948486 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-dgxcw" podStartSLOduration=135.948469439 podStartE2EDuration="2m15.948469439s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:31.945281616 +0000 UTC m=+154.688827867" watchObservedRunningTime="2025-12-10 09:36:31.948469439 +0000 UTC m=+154.692015690" Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.957772 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" event={"ID":"582cfff1-c3fb-462a-b80b-d14cd76a6c88","Type":"ContainerStarted","Data":"59c199a7e5c66f0eecb9c7dfbecd1207872ef4b40c2edf337a4f868a5b18f533"} Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.957815 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" event={"ID":"582cfff1-c3fb-462a-b80b-d14cd76a6c88","Type":"ContainerStarted","Data":"037594e817f2e689be9da4ec5055b6d37dffe12826480f6808acafa1801e3023"} Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.971248 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" event={"ID":"8a456b48-1823-4211-b1f9-cf64a5645e6a","Type":"ContainerStarted","Data":"95e4067f982b11a61d3c14ee73d23f54fb48d933938348346c695eff7c428b4c"} Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.973681 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:31 crc kubenswrapper[4715]: E1210 09:36:31.974117 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.474094337 +0000 UTC m=+155.217640588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.976725 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" event={"ID":"53adb33e-d25b-412a-be71-82e3e6380a3e","Type":"ContainerStarted","Data":"c9b75563205f4284624a3fb168bdf6542a63c628e5d218228d360d69218c26d3"} Dec 10 09:36:31 crc kubenswrapper[4715]: I1210 09:36:31.982577 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" event={"ID":"b06279e0-9b31-40fb-ac5b-d7fd85abd9c8","Type":"ContainerStarted","Data":"2d027ef2927bc6ed63ab7a7ad042cf14467a78d35ec11436becb22909b276d24"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.008842 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5" event={"ID":"4435aa00-3653-4733-90ef-86f45701c917","Type":"ContainerStarted","Data":"4c3a5063b1c55dd6cc9744665df121c3177c187de8dc6d2132ce3d5489ac2b2e"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.034689 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" event={"ID":"412a7188-42b3-4f08-affa-f4df0c076e46","Type":"ContainerStarted","Data":"cdb822f7c3aae084ff61f55f3cd9181632cd4bdb19da79c5752a0386b54292ed"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.034728 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p2967" podStartSLOduration=136.034712026 podStartE2EDuration="2m16.034712026s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:31.9906332 +0000 UTC m=+154.734179451" watchObservedRunningTime="2025-12-10 09:36:32.034712026 +0000 UTC m=+154.778258267" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.036943 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" podStartSLOduration=136.036927141 podStartE2EDuration="2m16.036927141s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.034360336 +0000 UTC m=+154.777906587" watchObservedRunningTime="2025-12-10 09:36:32.036927141 +0000 UTC m=+154.780473382" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.047967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" event={"ID":"500b95ee-e25a-4aff-97f0-f8af3b1f8831","Type":"ContainerStarted","Data":"72424c5c6fc63249e0990269e51c5639762429885eba9b47222e1ca4589d3684"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.061081 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6njjb" event={"ID":"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79","Type":"ContainerStarted","Data":"915c5ff07f9a08e789ddd19ceba5edf5ceb7c9b4debbdbeaf2356498438140a4"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.067791 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7bhpw" podStartSLOduration=136.067774141 podStartE2EDuration="2m16.067774141s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.061503898 +0000 UTC m=+154.805050149" watchObservedRunningTime="2025-12-10 09:36:32.067774141 +0000 UTC m=+154.811320392" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.077367 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:32 crc kubenswrapper[4715]: E1210 09:36:32.083326 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.583311865 +0000 UTC m=+155.326858116 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.089697 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" event={"ID":"136c04d9-6eec-4412-91c3-0c426e0b7746","Type":"ContainerStarted","Data":"d244fd9d818fca4dd81472e25236a43a4d20fcd9cfca39e9d012b36b1ea20112"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.090309 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.107180 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" event={"ID":"d347b264-47d4-4258-9473-ab7fb58a4cfa","Type":"ContainerStarted","Data":"a9bcef8c7a058f99a0bd9feaa72efcd7cbf56941d75f8161321f749e8a28730f"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.147226 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" event={"ID":"851585ad-0aba-46c3-8161-b329fb65655b","Type":"ContainerStarted","Data":"15c526d71716191f752a89ac6e69bb83825cf4c40151412fe083d6907488e4e5"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.149379 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h4crp" podStartSLOduration=136.149363583 podStartE2EDuration="2m16.149363583s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.103600357 +0000 UTC m=+154.847146608" watchObservedRunningTime="2025-12-10 09:36:32.149363583 +0000 UTC m=+154.892909834" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.178641 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:32 crc kubenswrapper[4715]: E1210 09:36:32.178738 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.67871544 +0000 UTC m=+155.422261681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.179163 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:32 crc kubenswrapper[4715]: E1210 09:36:32.181536 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.681518772 +0000 UTC m=+155.425065023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.182130 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" event={"ID":"29764aad-227f-400b-9ebd-703c0e69f519","Type":"ContainerStarted","Data":"aef07f41a7025e24dcf2f2c5cce893023bc8e93e8c7f4819fbfcdbd814f13ac3"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.201398 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-q6645" podStartSLOduration=136.201376751 podStartE2EDuration="2m16.201376751s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.151196146 +0000 UTC m=+154.894742397" watchObservedRunningTime="2025-12-10 09:36:32.201376751 +0000 UTC m=+154.944923002" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.201554 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-74xmx" podStartSLOduration=136.201549326 podStartE2EDuration="2m16.201549326s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.198408685 +0000 UTC m=+154.941954946" watchObservedRunningTime="2025-12-10 09:36:32.201549326 +0000 UTC m=+154.945095577" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.229256 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" event={"ID":"1fb0006d-63be-4593-a64e-9af6225de71a","Type":"ContainerStarted","Data":"dd4129519364ce9beba76c3f70ee6ec8c3425547850dc438c0cbb900109760a3"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.230464 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.241339 4715 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w76w7 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.241404 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" podUID="1fb0006d-63be-4593-a64e-9af6225de71a" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.258584 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hg2rw" event={"ID":"8c174d6b-eb6a-461c-a58b-7f497a5118e6","Type":"ContainerStarted","Data":"893a39c7356c1fd0f63960c31768c70ea500ccda8709ebb803ace6c8604ff294"} Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.264172 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dnj8g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.264248 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" podUID="f37a1d66-5d7d-4d58-9f23-401de6615848" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.264317 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.264362 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.295894 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:32 crc kubenswrapper[4715]: E1210 09:36:32.296205 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.796187709 +0000 UTC m=+155.539733960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.300039 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8r4m5" podStartSLOduration=136.30001726 podStartE2EDuration="2m16.30001726s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.254006037 +0000 UTC m=+154.997552288" watchObservedRunningTime="2025-12-10 09:36:32.30001726 +0000 UTC m=+155.043563511" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.351230 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25cwg" podStartSLOduration=136.351210205 podStartE2EDuration="2m16.351210205s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.295292242 +0000 UTC m=+155.038838503" watchObservedRunningTime="2025-12-10 09:36:32.351210205 +0000 UTC m=+155.094756456" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.380954 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" podStartSLOduration=135.380895771 podStartE2EDuration="2m15.380895771s" podCreationTimestamp="2025-12-10 09:34:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.352213044 +0000 UTC m=+155.095759295" watchObservedRunningTime="2025-12-10 09:36:32.380895771 +0000 UTC m=+155.124442012" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.383158 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" podStartSLOduration=136.383146827 podStartE2EDuration="2m16.383146827s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.380538291 +0000 UTC m=+155.124084542" watchObservedRunningTime="2025-12-10 09:36:32.383146827 +0000 UTC m=+155.126693068" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.398345 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:32 crc kubenswrapper[4715]: E1210 09:36:32.399797 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:32.899784153 +0000 UTC m=+155.643330404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.498965 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-8wtv4" podStartSLOduration=136.498948047 podStartE2EDuration="2m16.498948047s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.420047104 +0000 UTC m=+155.163593355" watchObservedRunningTime="2025-12-10 09:36:32.498948047 +0000 UTC m=+155.242494288" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.500678 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:32 crc kubenswrapper[4715]: E1210 09:36:32.501258 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.001234194 +0000 UTC m=+155.744780445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.501555 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-q5bbl" podStartSLOduration=136.501538803 podStartE2EDuration="2m16.501538803s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.498445553 +0000 UTC m=+155.241991804" watchObservedRunningTime="2025-12-10 09:36:32.501538803 +0000 UTC m=+155.245085054" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.557296 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-hg2rw" podStartSLOduration=10.55726499 podStartE2EDuration="10.55726499s" podCreationTimestamp="2025-12-10 09:36:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.555930261 +0000 UTC m=+155.299476512" watchObservedRunningTime="2025-12-10 09:36:32.55726499 +0000 UTC m=+155.300811241" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.602811 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:32 crc kubenswrapper[4715]: E1210 09:36:32.603254 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.103233231 +0000 UTC m=+155.846779482 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.644599 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:32 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:32 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:32 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.644667 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.704531 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:32 crc kubenswrapper[4715]: E1210 09:36:32.705183 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.205137556 +0000 UTC m=+155.948683807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.805807 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:32 crc kubenswrapper[4715]: E1210 09:36:32.806415 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.306391841 +0000 UTC m=+156.049938092 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:32 crc kubenswrapper[4715]: I1210 09:36:32.906285 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:32 crc kubenswrapper[4715]: E1210 09:36:32.906631 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.406610927 +0000 UTC m=+156.150157178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.008240 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.008728 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.508712117 +0000 UTC m=+156.252258368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.108639 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.108979 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.608945183 +0000 UTC m=+156.352491444 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.109173 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.109486 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.609471358 +0000 UTC m=+156.353017609 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.205381 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nxdbh" Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.210102 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.210303 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.710273711 +0000 UTC m=+156.453819962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.275004 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6njjb" event={"ID":"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79","Type":"ContainerStarted","Data":"6f742b95919b4676a759aba7788ef70d3fcced77a5f3b577f613d22606faaef1"} Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.275093 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6njjb" event={"ID":"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79","Type":"ContainerStarted","Data":"3edbc7258c22cbeecbe9258ff8af9172bd2224629195e41ebbd12584d87cc3a5"} Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.276928 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.277606 4715 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-tsqzx container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.277657 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" podUID="136c04d9-6eec-4412-91c3-0c426e0b7746" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.278372 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.278418 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.287462 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.291233 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w76w7" podStartSLOduration=136.291212053 podStartE2EDuration="2m16.291212053s" podCreationTimestamp="2025-12-10 09:34:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:32.586858083 +0000 UTC m=+155.330404324" watchObservedRunningTime="2025-12-10 09:36:33.291212053 +0000 UTC m=+156.034758294" Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.312676 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.315375 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.815354128 +0000 UTC m=+156.558900379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.337755 4715 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.419030 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.419194 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.919164888 +0000 UTC m=+156.662711139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.419256 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.419558 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:33.919543639 +0000 UTC m=+156.663089890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.520313 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.521070 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:34.021033171 +0000 UTC m=+156.764579422 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.621999 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.622596 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:34.122571235 +0000 UTC m=+156.866117486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.642853 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:33 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:33 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:33 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.642978 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.723706 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.724284 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 09:36:34.224263804 +0000 UTC m=+156.967810055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.825462 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:33 crc kubenswrapper[4715]: E1210 09:36:33.825885 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 09:36:34.32586507 +0000 UTC m=+157.069411321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nllws" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.828904 4715 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-10T09:36:33.337794423Z","Handler":null,"Name":""} Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.870597 4715 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.870648 4715 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.926090 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 09:36:33 crc kubenswrapper[4715]: I1210 09:36:33.978515 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.028007 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.147369 4715 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.147420 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.244499 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.244540 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.251659 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.252590 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.254498 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fx8wx"] Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.255882 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.256105 4715 patch_prober.go:28] interesting pod/console-f9d7485db-zphsm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.256142 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zphsm" podUID="5fbc3706-b2a3-48aa-a950-dc403907fa02" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.261802 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-7lwft" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.262702 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.265151 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fx8wx"] Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.271837 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nllws\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.307428 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6njjb" event={"ID":"429ceb66-19e6-4fa6-bd59-dc06aa7c3d79","Type":"ContainerStarted","Data":"6e8d2c9e9aed4ff365a1c33bc4a0ee6fcf9e896d00dc8e79b3f76367b8e2d3ff"} Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.332986 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-6njjb" podStartSLOduration=12.332970992 podStartE2EDuration="12.332970992s" podCreationTimestamp="2025-12-10 09:36:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:34.331373985 +0000 UTC m=+157.074920236" watchObservedRunningTime="2025-12-10 09:36:34.332970992 +0000 UTC m=+157.076517243" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.368899 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.368974 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.369506 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.369559 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.436419 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-utilities\") pod \"certified-operators-fx8wx\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.437239 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgb4r\" (UniqueName: \"kubernetes.io/projected/078de70a-e498-487c-81a0-e7e3532e7c48-kube-api-access-jgb4r\") pod \"certified-operators-fx8wx\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.437439 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-catalog-content\") pod \"certified-operators-fx8wx\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.436875 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vq4cc"] Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.438446 4715 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-tsqzx container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.438556 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" podUID="136c04d9-6eec-4412-91c3-0c426e0b7746" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.439337 4715 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-tsqzx container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.439411 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" podUID="136c04d9-6eec-4412-91c3-0c426e0b7746" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.440536 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.446516 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.465514 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vq4cc"] Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.538265 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-utilities\") pod \"certified-operators-fx8wx\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.538705 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgb4r\" (UniqueName: \"kubernetes.io/projected/078de70a-e498-487c-81a0-e7e3532e7c48-kube-api-access-jgb4r\") pod \"certified-operators-fx8wx\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.538842 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-utilities\") pod \"community-operators-vq4cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.538985 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-catalog-content\") pod \"certified-operators-fx8wx\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.538854 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-utilities\") pod \"certified-operators-fx8wx\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.539127 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbjkm\" (UniqueName: \"kubernetes.io/projected/fa81c629-c83d-444d-a113-de3681b8f0cc-kube-api-access-lbjkm\") pod \"community-operators-vq4cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.539216 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-catalog-content\") pod \"community-operators-vq4cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.539429 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-catalog-content\") pod \"certified-operators-fx8wx\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.575232 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.591160 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgb4r\" (UniqueName: \"kubernetes.io/projected/078de70a-e498-487c-81a0-e7e3532e7c48-kube-api-access-jgb4r\") pod \"certified-operators-fx8wx\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.638803 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.640208 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-utilities\") pod \"community-operators-vq4cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.640263 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbjkm\" (UniqueName: \"kubernetes.io/projected/fa81c629-c83d-444d-a113-de3681b8f0cc-kube-api-access-lbjkm\") pod \"community-operators-vq4cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.640292 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-catalog-content\") pod \"community-operators-vq4cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.641200 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-catalog-content\") pod \"community-operators-vq4cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.641481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-utilities\") pod \"community-operators-vq4cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.655293 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-tsqzx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.663848 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9n9gx"] Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.665380 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.684012 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9n9gx"] Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.699853 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbjkm\" (UniqueName: \"kubernetes.io/projected/fa81c629-c83d-444d-a113-de3681b8f0cc-kube-api-access-lbjkm\") pod \"community-operators-vq4cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.726048 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:34 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:34 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:34 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.726115 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.751862 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-utilities\") pod \"certified-operators-9n9gx\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.751957 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-catalog-content\") pod \"certified-operators-9n9gx\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.751983 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btvn2\" (UniqueName: \"kubernetes.io/projected/efadbc66-5bb8-46c9-88cf-482defd875bd-kube-api-access-btvn2\") pod \"certified-operators-9n9gx\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.756850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.831419 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6fxk6"] Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.832374 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.853697 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-utilities\") pod \"certified-operators-9n9gx\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.853982 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-catalog-content\") pod \"certified-operators-9n9gx\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.854103 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-utilities\") pod \"community-operators-6fxk6\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.854184 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btvn2\" (UniqueName: \"kubernetes.io/projected/efadbc66-5bb8-46c9-88cf-482defd875bd-kube-api-access-btvn2\") pod \"certified-operators-9n9gx\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.854262 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t6ln\" (UniqueName: \"kubernetes.io/projected/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-kube-api-access-5t6ln\") pod \"community-operators-6fxk6\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.854344 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-catalog-content\") pod \"community-operators-6fxk6\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.854873 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-utilities\") pod \"certified-operators-9n9gx\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.855190 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-catalog-content\") pod \"certified-operators-9n9gx\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.863963 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6fxk6"] Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.883501 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.883605 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btvn2\" (UniqueName: \"kubernetes.io/projected/efadbc66-5bb8-46c9-88cf-482defd875bd-kube-api-access-btvn2\") pod \"certified-operators-9n9gx\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.956893 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-utilities\") pod \"community-operators-6fxk6\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.957355 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t6ln\" (UniqueName: \"kubernetes.io/projected/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-kube-api-access-5t6ln\") pod \"community-operators-6fxk6\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.957375 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-catalog-content\") pod \"community-operators-6fxk6\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.959644 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-catalog-content\") pod \"community-operators-6fxk6\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.959855 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-utilities\") pod \"community-operators-6fxk6\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.983369 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.989384 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.992632 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.992903 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 10 09:36:34 crc kubenswrapper[4715]: I1210 09:36:34.998496 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.046254 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.046295 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.056786 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t6ln\" (UniqueName: \"kubernetes.io/projected/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-kube-api-access-5t6ln\") pod \"community-operators-6fxk6\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.059134 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e04a0eeb-8cc5-49a2-a703-557a61d5028b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.059180 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e04a0eeb-8cc5-49a2-a703-557a61d5028b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.073206 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.074524 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.160579 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e04a0eeb-8cc5-49a2-a703-557a61d5028b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.160699 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e04a0eeb-8cc5-49a2-a703-557a61d5028b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.160768 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e04a0eeb-8cc5-49a2-a703-557a61d5028b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.162818 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.175769 4715 patch_prober.go:28] interesting pod/apiserver-76f77b778f-p5sqh container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]log ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]etcd ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]poststarthook/generic-apiserver-start-informers ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]poststarthook/max-in-flight-filter ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 10 09:36:35 crc kubenswrapper[4715]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 10 09:36:35 crc kubenswrapper[4715]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 10 09:36:35 crc kubenswrapper[4715]: [+]poststarthook/project.openshift.io-projectcache ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]poststarthook/openshift.io-startinformers ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 10 09:36:35 crc kubenswrapper[4715]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 10 09:36:35 crc kubenswrapper[4715]: livez check failed Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.175825 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" podUID="53adb33e-d25b-412a-be71-82e3e6380a3e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.204959 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e04a0eeb-8cc5-49a2-a703-557a61d5028b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.411214 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.457257 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7g7c9" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.461751 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.533992 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nllws"] Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.616723 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.651188 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:35 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:35 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:35 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.651245 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.702410 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vq4cc"] Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.770691 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9n9gx"] Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.898386 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fx8wx"] Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.916803 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6fxk6"] Dec 10 09:36:35 crc kubenswrapper[4715]: I1210 09:36:35.922523 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 09:36:35 crc kubenswrapper[4715]: W1210 09:36:35.932185 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9083d000_4c4d_4e48_82d7_1f7a1ec885b0.slice/crio-cf4973b416d6f947364eb825c31b4e6dd0a5ca17b165b4ea0f9dbfd420b3f052 WatchSource:0}: Error finding container cf4973b416d6f947364eb825c31b4e6dd0a5ca17b165b4ea0f9dbfd420b3f052: Status 404 returned error can't find the container with id cf4973b416d6f947364eb825c31b4e6dd0a5ca17b165b4ea0f9dbfd420b3f052 Dec 10 09:36:35 crc kubenswrapper[4715]: W1210 09:36:35.933616 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode04a0eeb_8cc5_49a2_a703_557a61d5028b.slice/crio-9edcb795b712dcec584e990d6171a88b0979d3dc98c4563592b33726c7efda2b WatchSource:0}: Error finding container 9edcb795b712dcec584e990d6171a88b0979d3dc98c4563592b33726c7efda2b: Status 404 returned error can't find the container with id 9edcb795b712dcec584e990d6171a88b0979d3dc98c4563592b33726c7efda2b Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.428041 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8t5q2"] Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.429643 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.431779 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.437344 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fx8wx" event={"ID":"078de70a-e498-487c-81a0-e7e3532e7c48","Type":"ContainerStarted","Data":"1c89e1c0681d3392c630569a2783cc732274983b4b44a0088b0693a46706e275"} Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.442205 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9n9gx" event={"ID":"efadbc66-5bb8-46c9-88cf-482defd875bd","Type":"ContainerStarted","Data":"bc0d4f8c240d9c68ca7b68ce9709e02f5c3bbaac9192e93c89679ccdc74e40d0"} Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.442556 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8t5q2"] Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.443118 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fxk6" event={"ID":"9083d000-4c4d-4e48-82d7-1f7a1ec885b0","Type":"ContainerStarted","Data":"cf4973b416d6f947364eb825c31b4e6dd0a5ca17b165b4ea0f9dbfd420b3f052"} Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.443503 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p727r\" (UniqueName: \"kubernetes.io/projected/676753c5-d951-4722-a5ef-82e3227e0c7e-kube-api-access-p727r\") pod \"redhat-marketplace-8t5q2\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.443605 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-catalog-content\") pod \"redhat-marketplace-8t5q2\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.443644 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-utilities\") pod \"redhat-marketplace-8t5q2\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.444005 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" event={"ID":"3dcd1e77-33f0-4a5b-95c5-7454b648db5c","Type":"ContainerStarted","Data":"16802a7acc9d1e69a52542c48f0314ace074a7d9a32658eae1ee4593ca3f4682"} Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.445027 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq4cc" event={"ID":"fa81c629-c83d-444d-a113-de3681b8f0cc","Type":"ContainerStarted","Data":"92be6d23c3789901906037041ddc11a0130c3415dd69608d9b8e3311fe81be39"} Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.445996 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e04a0eeb-8cc5-49a2-a703-557a61d5028b","Type":"ContainerStarted","Data":"9edcb795b712dcec584e990d6171a88b0979d3dc98c4563592b33726c7efda2b"} Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.545258 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p727r\" (UniqueName: \"kubernetes.io/projected/676753c5-d951-4722-a5ef-82e3227e0c7e-kube-api-access-p727r\") pod \"redhat-marketplace-8t5q2\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.545331 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-catalog-content\") pod \"redhat-marketplace-8t5q2\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.545353 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-utilities\") pod \"redhat-marketplace-8t5q2\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.545877 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-utilities\") pod \"redhat-marketplace-8t5q2\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.545993 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-catalog-content\") pod \"redhat-marketplace-8t5q2\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.573658 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p727r\" (UniqueName: \"kubernetes.io/projected/676753c5-d951-4722-a5ef-82e3227e0c7e-kube-api-access-p727r\") pod \"redhat-marketplace-8t5q2\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.643030 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:36 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:36 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:36 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.643163 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.757333 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.841072 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xsbgs"] Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.842344 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.852019 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsbgs"] Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.857816 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-catalog-content\") pod \"redhat-marketplace-xsbgs\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.858021 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98xzg\" (UniqueName: \"kubernetes.io/projected/88de80a3-3467-4439-82ac-3a26dd5073b4-kube-api-access-98xzg\") pod \"redhat-marketplace-xsbgs\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.858105 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-utilities\") pod \"redhat-marketplace-xsbgs\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.958895 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98xzg\" (UniqueName: \"kubernetes.io/projected/88de80a3-3467-4439-82ac-3a26dd5073b4-kube-api-access-98xzg\") pod \"redhat-marketplace-xsbgs\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.959013 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-utilities\") pod \"redhat-marketplace-xsbgs\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.959062 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-catalog-content\") pod \"redhat-marketplace-xsbgs\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.959604 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-catalog-content\") pod \"redhat-marketplace-xsbgs\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.959648 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-utilities\") pod \"redhat-marketplace-xsbgs\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:36 crc kubenswrapper[4715]: I1210 09:36:36.982478 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98xzg\" (UniqueName: \"kubernetes.io/projected/88de80a3-3467-4439-82ac-3a26dd5073b4-kube-api-access-98xzg\") pod \"redhat-marketplace-xsbgs\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.165424 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.294044 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8t5q2"] Dec 10 09:36:37 crc kubenswrapper[4715]: W1210 09:36:37.303706 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod676753c5_d951_4722_a5ef_82e3227e0c7e.slice/crio-77f7460c587d9dee9029c21e47756c323a831ed5bf2bb0fbec71bff2c409cbfe WatchSource:0}: Error finding container 77f7460c587d9dee9029c21e47756c323a831ed5bf2bb0fbec71bff2c409cbfe: Status 404 returned error can't find the container with id 77f7460c587d9dee9029c21e47756c323a831ed5bf2bb0fbec71bff2c409cbfe Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.377567 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.458445 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fxk6" event={"ID":"9083d000-4c4d-4e48-82d7-1f7a1ec885b0","Type":"ContainerStarted","Data":"babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9"} Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.458752 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jrl25"] Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.459935 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.462199 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" event={"ID":"3dcd1e77-33f0-4a5b-95c5-7454b648db5c","Type":"ContainerStarted","Data":"d9d6ae89e361ab4c8ae2a25bba5e37f34eea0ca6fb9cc521576d331baddd5bb0"} Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.462449 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.463791 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.464475 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t5q2" event={"ID":"676753c5-d951-4722-a5ef-82e3227e0c7e","Type":"ContainerStarted","Data":"77f7460c587d9dee9029c21e47756c323a831ed5bf2bb0fbec71bff2c409cbfe"} Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.465676 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsbgs"] Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.466020 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq4cc" event={"ID":"fa81c629-c83d-444d-a113-de3681b8f0cc","Type":"ContainerStarted","Data":"baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da"} Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.469261 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jrl25"] Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.477136 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zp86\" (UniqueName: \"kubernetes.io/projected/6d6da964-23be-49cc-8555-97edb12b5f28-kube-api-access-6zp86\") pod \"redhat-operators-jrl25\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.477213 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-catalog-content\") pod \"redhat-operators-jrl25\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.477322 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-utilities\") pod \"redhat-operators-jrl25\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.484890 4715 generic.go:334] "Generic (PLEG): container finished" podID="efb368e0-756b-4323-8ca2-0852ff385eb4" containerID="ba038ccc002ae5e69cf5603e7af827400c078a0a8a5db976d6d1797f0bbdaa48" exitCode=0 Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.484967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" event={"ID":"efb368e0-756b-4323-8ca2-0852ff385eb4","Type":"ContainerDied","Data":"ba038ccc002ae5e69cf5603e7af827400c078a0a8a5db976d6d1797f0bbdaa48"} Dec 10 09:36:37 crc kubenswrapper[4715]: W1210 09:36:37.503989 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88de80a3_3467_4439_82ac_3a26dd5073b4.slice/crio-ccfbdaf22d9428c76a62c87d47bc1278ceb0f07881eed905270f7873f7f6cdec WatchSource:0}: Error finding container ccfbdaf22d9428c76a62c87d47bc1278ceb0f07881eed905270f7873f7f6cdec: Status 404 returned error can't find the container with id ccfbdaf22d9428c76a62c87d47bc1278ceb0f07881eed905270f7873f7f6cdec Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.509636 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e04a0eeb-8cc5-49a2-a703-557a61d5028b","Type":"ContainerStarted","Data":"ed48be104ab00652e87b13e73a7f43d851df31bb411389e187a278c04398291d"} Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.537246 4715 generic.go:334] "Generic (PLEG): container finished" podID="078de70a-e498-487c-81a0-e7e3532e7c48" containerID="e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d" exitCode=0 Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.537349 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fx8wx" event={"ID":"078de70a-e498-487c-81a0-e7e3532e7c48","Type":"ContainerDied","Data":"e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d"} Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.564330 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" podStartSLOduration=141.564297802 podStartE2EDuration="2m21.564297802s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:37.548220023 +0000 UTC m=+160.291766274" watchObservedRunningTime="2025-12-10 09:36:37.564297802 +0000 UTC m=+160.307844123" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.576464 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.584230 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-utilities\") pod \"redhat-operators-jrl25\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.584404 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zp86\" (UniqueName: \"kubernetes.io/projected/6d6da964-23be-49cc-8555-97edb12b5f28-kube-api-access-6zp86\") pod \"redhat-operators-jrl25\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.584462 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-catalog-content\") pod \"redhat-operators-jrl25\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.585592 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-utilities\") pod \"redhat-operators-jrl25\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.617705 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-catalog-content\") pod \"redhat-operators-jrl25\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.635381 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.635349036 podStartE2EDuration="3.635349036s" podCreationTimestamp="2025-12-10 09:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:37.627802366 +0000 UTC m=+160.371348617" watchObservedRunningTime="2025-12-10 09:36:37.635349036 +0000 UTC m=+160.378895287" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.641086 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xn5c9"] Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.643632 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:37 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:37 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:37 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.643726 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.645531 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.647389 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xn5c9"] Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.655307 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zp86\" (UniqueName: \"kubernetes.io/projected/6d6da964-23be-49cc-8555-97edb12b5f28-kube-api-access-6zp86\") pod \"redhat-operators-jrl25\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.755036 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.786745 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-utilities\") pod \"redhat-operators-xn5c9\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.786814 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42w9r\" (UniqueName: \"kubernetes.io/projected/8c078273-d5a5-4371-8ad5-091114f61311-kube-api-access-42w9r\") pod \"redhat-operators-xn5c9\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.786863 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-catalog-content\") pod \"redhat-operators-xn5c9\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.887533 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-catalog-content\") pod \"redhat-operators-xn5c9\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.887619 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-utilities\") pod \"redhat-operators-xn5c9\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.887656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42w9r\" (UniqueName: \"kubernetes.io/projected/8c078273-d5a5-4371-8ad5-091114f61311-kube-api-access-42w9r\") pod \"redhat-operators-xn5c9\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.888378 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-utilities\") pod \"redhat-operators-xn5c9\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.889575 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-catalog-content\") pod \"redhat-operators-xn5c9\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:37 crc kubenswrapper[4715]: I1210 09:36:37.910534 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42w9r\" (UniqueName: \"kubernetes.io/projected/8c078273-d5a5-4371-8ad5-091114f61311-kube-api-access-42w9r\") pod \"redhat-operators-xn5c9\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.027342 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jrl25"] Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.108600 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.396037 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xn5c9"] Dec 10 09:36:38 crc kubenswrapper[4715]: W1210 09:36:38.467053 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c078273_d5a5_4371_8ad5_091114f61311.slice/crio-89e3b97893a5169ad240698afc2fb189a4eb64404cdd443a15a9f6cfb10e9721 WatchSource:0}: Error finding container 89e3b97893a5169ad240698afc2fb189a4eb64404cdd443a15a9f6cfb10e9721: Status 404 returned error can't find the container with id 89e3b97893a5169ad240698afc2fb189a4eb64404cdd443a15a9f6cfb10e9721 Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.496981 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.515277 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f450d9cd-df1c-456e-861b-d91b2b683417-metrics-certs\") pod \"network-metrics-daemon-h5xdb\" (UID: \"f450d9cd-df1c-456e-861b-d91b2b683417\") " pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.573254 4715 generic.go:334] "Generic (PLEG): container finished" podID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerID="23b1f5fe5e829f796334cefb6d23e35fe4e33c7095fdd0759ac4f3ebb24047d0" exitCode=0 Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.573323 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9n9gx" event={"ID":"efadbc66-5bb8-46c9-88cf-482defd875bd","Type":"ContainerDied","Data":"23b1f5fe5e829f796334cefb6d23e35fe4e33c7095fdd0759ac4f3ebb24047d0"} Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.577945 4715 generic.go:334] "Generic (PLEG): container finished" podID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerID="babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9" exitCode=0 Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.578018 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fxk6" event={"ID":"9083d000-4c4d-4e48-82d7-1f7a1ec885b0","Type":"ContainerDied","Data":"babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9"} Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.583389 4715 generic.go:334] "Generic (PLEG): container finished" podID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerID="37f48a809b5c9f81628cd8ddfb6265ee7b5016a9f614653652482092bd2a0976" exitCode=0 Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.583440 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsbgs" event={"ID":"88de80a3-3467-4439-82ac-3a26dd5073b4","Type":"ContainerDied","Data":"37f48a809b5c9f81628cd8ddfb6265ee7b5016a9f614653652482092bd2a0976"} Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.583462 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsbgs" event={"ID":"88de80a3-3467-4439-82ac-3a26dd5073b4","Type":"ContainerStarted","Data":"ccfbdaf22d9428c76a62c87d47bc1278ceb0f07881eed905270f7873f7f6cdec"} Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.585699 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xn5c9" event={"ID":"8c078273-d5a5-4371-8ad5-091114f61311","Type":"ContainerStarted","Data":"89e3b97893a5169ad240698afc2fb189a4eb64404cdd443a15a9f6cfb10e9721"} Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.588352 4715 generic.go:334] "Generic (PLEG): container finished" podID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerID="cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c" exitCode=0 Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.588467 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t5q2" event={"ID":"676753c5-d951-4722-a5ef-82e3227e0c7e","Type":"ContainerDied","Data":"cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c"} Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.610254 4715 generic.go:334] "Generic (PLEG): container finished" podID="6d6da964-23be-49cc-8555-97edb12b5f28" containerID="97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e" exitCode=0 Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.610383 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrl25" event={"ID":"6d6da964-23be-49cc-8555-97edb12b5f28","Type":"ContainerDied","Data":"97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e"} Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.610415 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrl25" event={"ID":"6d6da964-23be-49cc-8555-97edb12b5f28","Type":"ContainerStarted","Data":"344c2e1defc32a94cd0552d88ecdd5a6cfd6cbf3486ea43452d5f19a85500103"} Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.615397 4715 generic.go:334] "Generic (PLEG): container finished" podID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerID="baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da" exitCode=0 Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.615575 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq4cc" event={"ID":"fa81c629-c83d-444d-a113-de3681b8f0cc","Type":"ContainerDied","Data":"baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da"} Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.620001 4715 generic.go:334] "Generic (PLEG): container finished" podID="e04a0eeb-8cc5-49a2-a703-557a61d5028b" containerID="ed48be104ab00652e87b13e73a7f43d851df31bb411389e187a278c04398291d" exitCode=0 Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.620125 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e04a0eeb-8cc5-49a2-a703-557a61d5028b","Type":"ContainerDied","Data":"ed48be104ab00652e87b13e73a7f43d851df31bb411389e187a278c04398291d"} Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.643154 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:38 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:38 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:38 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.643556 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:38 crc kubenswrapper[4715]: I1210 09:36:38.716673 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-h5xdb" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.084004 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.111516 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-h5xdb"] Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.116072 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efb368e0-756b-4323-8ca2-0852ff385eb4-config-volume\") pod \"efb368e0-756b-4323-8ca2-0852ff385eb4\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.116244 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/efb368e0-756b-4323-8ca2-0852ff385eb4-secret-volume\") pod \"efb368e0-756b-4323-8ca2-0852ff385eb4\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.116318 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc72z\" (UniqueName: \"kubernetes.io/projected/efb368e0-756b-4323-8ca2-0852ff385eb4-kube-api-access-kc72z\") pod \"efb368e0-756b-4323-8ca2-0852ff385eb4\" (UID: \"efb368e0-756b-4323-8ca2-0852ff385eb4\") " Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.117275 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb368e0-756b-4323-8ca2-0852ff385eb4-config-volume" (OuterVolumeSpecName: "config-volume") pod "efb368e0-756b-4323-8ca2-0852ff385eb4" (UID: "efb368e0-756b-4323-8ca2-0852ff385eb4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.124515 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efb368e0-756b-4323-8ca2-0852ff385eb4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "efb368e0-756b-4323-8ca2-0852ff385eb4" (UID: "efb368e0-756b-4323-8ca2-0852ff385eb4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.127311 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.128080 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efb368e0-756b-4323-8ca2-0852ff385eb4-kube-api-access-kc72z" (OuterVolumeSpecName: "kube-api-access-kc72z") pod "efb368e0-756b-4323-8ca2-0852ff385eb4" (UID: "efb368e0-756b-4323-8ca2-0852ff385eb4"). InnerVolumeSpecName "kube-api-access-kc72z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:36:39 crc kubenswrapper[4715]: E1210 09:36:39.128449 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efb368e0-756b-4323-8ca2-0852ff385eb4" containerName="collect-profiles" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.128468 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="efb368e0-756b-4323-8ca2-0852ff385eb4" containerName="collect-profiles" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.128593 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="efb368e0-756b-4323-8ca2-0852ff385eb4" containerName="collect-profiles" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.129127 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.131102 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.136747 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.158357 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.217535 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.217579 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.217810 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/efb368e0-756b-4323-8ca2-0852ff385eb4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.217836 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc72z\" (UniqueName: \"kubernetes.io/projected/efb368e0-756b-4323-8ca2-0852ff385eb4-kube-api-access-kc72z\") on node \"crc\" DevicePath \"\"" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.217847 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efb368e0-756b-4323-8ca2-0852ff385eb4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.249588 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.263315 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-p5sqh" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.319336 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.319402 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.321433 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.352703 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.466191 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.646729 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:39 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:39 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:39 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.647288 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.677416 4715 generic.go:334] "Generic (PLEG): container finished" podID="8c078273-d5a5-4371-8ad5-091114f61311" containerID="89e4fdc15f388f2a2a736ecc0736066e49d6ec159b3056e5d2c94321192cc005" exitCode=0 Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.678222 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xn5c9" event={"ID":"8c078273-d5a5-4371-8ad5-091114f61311","Type":"ContainerDied","Data":"89e4fdc15f388f2a2a736ecc0736066e49d6ec159b3056e5d2c94321192cc005"} Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.766121 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" event={"ID":"f450d9cd-df1c-456e-861b-d91b2b683417","Type":"ContainerStarted","Data":"7f6e75e4aa2b9ecb122cc4acbfc01b718c6905ef1a9743309ffdfde7ce1cda4f"} Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.766210 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" event={"ID":"f450d9cd-df1c-456e-861b-d91b2b683417","Type":"ContainerStarted","Data":"d704cb8aa80f3638d53898f2c9ccea7e0474dd2aa4e569e23ae7505c2bb51b97"} Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.779703 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.780381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975" event={"ID":"efb368e0-756b-4323-8ca2-0852ff385eb4","Type":"ContainerDied","Data":"3c5652c18d9c52ccd46ead026ad93beb3bbf377d4cb79748b16b6b1ddf36d4ac"} Dec 10 09:36:39 crc kubenswrapper[4715]: I1210 09:36:39.780471 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c5652c18d9c52ccd46ead026ad93beb3bbf377d4cb79748b16b6b1ddf36d4ac" Dec 10 09:36:40 crc kubenswrapper[4715]: I1210 09:36:40.824440 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:40 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:40 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:40 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:40 crc kubenswrapper[4715]: I1210 09:36:40.825048 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:40 crc kubenswrapper[4715]: I1210 09:36:40.833408 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-hg2rw" Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.007484 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.352444 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.496246 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kube-api-access\") pod \"e04a0eeb-8cc5-49a2-a703-557a61d5028b\" (UID: \"e04a0eeb-8cc5-49a2-a703-557a61d5028b\") " Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.496332 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kubelet-dir\") pod \"e04a0eeb-8cc5-49a2-a703-557a61d5028b\" (UID: \"e04a0eeb-8cc5-49a2-a703-557a61d5028b\") " Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.497199 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e04a0eeb-8cc5-49a2-a703-557a61d5028b" (UID: "e04a0eeb-8cc5-49a2-a703-557a61d5028b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.502944 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e04a0eeb-8cc5-49a2-a703-557a61d5028b" (UID: "e04a0eeb-8cc5-49a2-a703-557a61d5028b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.597939 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.597977 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e04a0eeb-8cc5-49a2-a703-557a61d5028b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.644463 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:41 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:41 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:41 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.644949 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.795046 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0","Type":"ContainerStarted","Data":"cd8124b24574db16b5aec1e2c86a0cf4c952a60019b5c5f9a92b9f6e0b8cba5e"} Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.798453 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e04a0eeb-8cc5-49a2-a703-557a61d5028b","Type":"ContainerDied","Data":"9edcb795b712dcec584e990d6171a88b0979d3dc98c4563592b33726c7efda2b"} Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.798481 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.798496 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9edcb795b712dcec584e990d6171a88b0979d3dc98c4563592b33726c7efda2b" Dec 10 09:36:41 crc kubenswrapper[4715]: I1210 09:36:41.805159 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-h5xdb" event={"ID":"f450d9cd-df1c-456e-861b-d91b2b683417","Type":"ContainerStarted","Data":"a9ca91f877dce71e05e7413b36fe612835e2b67065896c2fefd067f614722bec"} Dec 10 09:36:42 crc kubenswrapper[4715]: I1210 09:36:42.643637 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:42 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:42 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:42 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:42 crc kubenswrapper[4715]: I1210 09:36:42.644043 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:42 crc kubenswrapper[4715]: I1210 09:36:42.831537 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0","Type":"ContainerStarted","Data":"5eb3bd68499cc526657e4fda8030db46f36c1a3c928cbf23310dd3ca2500c5b2"} Dec 10 09:36:42 crc kubenswrapper[4715]: I1210 09:36:42.852864 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-h5xdb" podStartSLOduration=146.85282911 podStartE2EDuration="2m26.85282911s" podCreationTimestamp="2025-12-10 09:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:41.84285671 +0000 UTC m=+164.586402971" watchObservedRunningTime="2025-12-10 09:36:42.85282911 +0000 UTC m=+165.596375361" Dec 10 09:36:42 crc kubenswrapper[4715]: I1210 09:36:42.858274 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.8582159369999998 podStartE2EDuration="3.858215937s" podCreationTimestamp="2025-12-10 09:36:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:36:42.852442529 +0000 UTC m=+165.595988790" watchObservedRunningTime="2025-12-10 09:36:42.858215937 +0000 UTC m=+165.601762208" Dec 10 09:36:43 crc kubenswrapper[4715]: I1210 09:36:43.642791 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:43 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:43 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:43 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:43 crc kubenswrapper[4715]: I1210 09:36:43.643108 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:43 crc kubenswrapper[4715]: I1210 09:36:43.868240 4715 generic.go:334] "Generic (PLEG): container finished" podID="b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0" containerID="5eb3bd68499cc526657e4fda8030db46f36c1a3c928cbf23310dd3ca2500c5b2" exitCode=0 Dec 10 09:36:43 crc kubenswrapper[4715]: I1210 09:36:43.868294 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0","Type":"ContainerDied","Data":"5eb3bd68499cc526657e4fda8030db46f36c1a3c928cbf23310dd3ca2500c5b2"} Dec 10 09:36:44 crc kubenswrapper[4715]: I1210 09:36:44.242002 4715 patch_prober.go:28] interesting pod/console-f9d7485db-zphsm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 10 09:36:44 crc kubenswrapper[4715]: I1210 09:36:44.242062 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zphsm" podUID="5fbc3706-b2a3-48aa-a950-dc403907fa02" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 10 09:36:44 crc kubenswrapper[4715]: I1210 09:36:44.371360 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:36:44 crc kubenswrapper[4715]: I1210 09:36:44.371613 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:36:44 crc kubenswrapper[4715]: I1210 09:36:44.371677 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:36:44 crc kubenswrapper[4715]: I1210 09:36:44.371428 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:36:44 crc kubenswrapper[4715]: I1210 09:36:44.641813 4715 patch_prober.go:28] interesting pod/router-default-5444994796-72qbm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 09:36:44 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 10 09:36:44 crc kubenswrapper[4715]: [+]process-running ok Dec 10 09:36:44 crc kubenswrapper[4715]: healthz check failed Dec 10 09:36:44 crc kubenswrapper[4715]: I1210 09:36:44.641906 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-72qbm" podUID="fc7f7af1-1312-42c8-bc68-4bb42832b42d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 09:36:45 crc kubenswrapper[4715]: I1210 09:36:45.643614 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:45 crc kubenswrapper[4715]: I1210 09:36:45.650760 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-72qbm" Dec 10 09:36:47 crc kubenswrapper[4715]: I1210 09:36:47.714548 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:36:47 crc kubenswrapper[4715]: I1210 09:36:47.714620 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.246313 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.250696 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.368575 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.369297 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.369344 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-b9k84" Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.368646 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.369646 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.369961 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"77ec3149beda57df0ce8099444cc73647c2202b18cf33a8722bf5c09999040df"} pod="openshift-console/downloads-7954f5f757-b9k84" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.370036 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" containerID="cri-o://77ec3149beda57df0ce8099444cc73647c2202b18cf33a8722bf5c09999040df" gracePeriod=2 Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.370188 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.370217 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:36:54 crc kubenswrapper[4715]: I1210 09:36:54.587552 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:36:57 crc kubenswrapper[4715]: I1210 09:36:57.072099 4715 generic.go:334] "Generic (PLEG): container finished" podID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerID="77ec3149beda57df0ce8099444cc73647c2202b18cf33a8722bf5c09999040df" exitCode=0 Dec 10 09:36:57 crc kubenswrapper[4715]: I1210 09:36:57.072196 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b9k84" event={"ID":"55cadec6-a68b-4d03-abc0-d9c628c3e4d0","Type":"ContainerDied","Data":"77ec3149beda57df0ce8099444cc73647c2202b18cf33a8722bf5c09999040df"} Dec 10 09:37:04 crc kubenswrapper[4715]: I1210 09:37:04.059390 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 09:37:04 crc kubenswrapper[4715]: I1210 09:37:04.115012 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-lzmqn" Dec 10 09:37:04 crc kubenswrapper[4715]: I1210 09:37:04.369136 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:37:04 crc kubenswrapper[4715]: I1210 09:37:04.369204 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:37:09 crc kubenswrapper[4715]: I1210 09:37:09.399944 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 09:37:09 crc kubenswrapper[4715]: I1210 09:37:09.410149 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kubelet-dir\") pod \"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0\" (UID: \"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0\") " Dec 10 09:37:09 crc kubenswrapper[4715]: I1210 09:37:09.410235 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kube-api-access\") pod \"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0\" (UID: \"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0\") " Dec 10 09:37:09 crc kubenswrapper[4715]: I1210 09:37:09.410590 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0" (UID: "b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:37:09 crc kubenswrapper[4715]: I1210 09:37:09.421739 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0" (UID: "b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:37:09 crc kubenswrapper[4715]: I1210 09:37:09.511652 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:37:09 crc kubenswrapper[4715]: I1210 09:37:09.511685 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 09:37:10 crc kubenswrapper[4715]: I1210 09:37:10.287719 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0","Type":"ContainerDied","Data":"cd8124b24574db16b5aec1e2c86a0cf4c952a60019b5c5f9a92b9f6e0b8cba5e"} Dec 10 09:37:10 crc kubenswrapper[4715]: I1210 09:37:10.287760 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd8124b24574db16b5aec1e2c86a0cf4c952a60019b5c5f9a92b9f6e0b8cba5e" Dec 10 09:37:10 crc kubenswrapper[4715]: I1210 09:37:10.287811 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.711077 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 09:37:12 crc kubenswrapper[4715]: E1210 09:37:12.711974 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0" containerName="pruner" Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.711993 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0" containerName="pruner" Dec 10 09:37:12 crc kubenswrapper[4715]: E1210 09:37:12.712036 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e04a0eeb-8cc5-49a2-a703-557a61d5028b" containerName="pruner" Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.712045 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e04a0eeb-8cc5-49a2-a703-557a61d5028b" containerName="pruner" Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.712313 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e04a0eeb-8cc5-49a2-a703-557a61d5028b" containerName="pruner" Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.712331 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6f6c8b0-c7e8-4c3d-b231-66b7c36a44f0" containerName="pruner" Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.717445 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.719263 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.839020 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.839217 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.943469 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c19345d5-5575-4321-a37e-045244edec62-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c19345d5-5575-4321-a37e-045244edec62\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 09:37:12 crc kubenswrapper[4715]: I1210 09:37:12.943533 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c19345d5-5575-4321-a37e-045244edec62-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c19345d5-5575-4321-a37e-045244edec62\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 09:37:13 crc kubenswrapper[4715]: I1210 09:37:13.044853 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c19345d5-5575-4321-a37e-045244edec62-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c19345d5-5575-4321-a37e-045244edec62\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 09:37:13 crc kubenswrapper[4715]: I1210 09:37:13.044950 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c19345d5-5575-4321-a37e-045244edec62-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c19345d5-5575-4321-a37e-045244edec62\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 09:37:13 crc kubenswrapper[4715]: I1210 09:37:13.044998 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c19345d5-5575-4321-a37e-045244edec62-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c19345d5-5575-4321-a37e-045244edec62\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 09:37:13 crc kubenswrapper[4715]: I1210 09:37:13.064328 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c19345d5-5575-4321-a37e-045244edec62-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c19345d5-5575-4321-a37e-045244edec62\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 09:37:13 crc kubenswrapper[4715]: I1210 09:37:13.167861 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 09:37:14 crc kubenswrapper[4715]: I1210 09:37:14.368767 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:37:14 crc kubenswrapper[4715]: I1210 09:37:14.368824 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:37:17 crc kubenswrapper[4715]: I1210 09:37:17.714168 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:37:17 crc kubenswrapper[4715]: I1210 09:37:17.714529 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:37:17 crc kubenswrapper[4715]: I1210 09:37:17.714580 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:37:17 crc kubenswrapper[4715]: I1210 09:37:17.716087 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 09:37:17 crc kubenswrapper[4715]: I1210 09:37:17.716173 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5" gracePeriod=600 Dec 10 09:37:17 crc kubenswrapper[4715]: I1210 09:37:17.905103 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 09:37:17 crc kubenswrapper[4715]: I1210 09:37:17.906162 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:17 crc kubenswrapper[4715]: I1210 09:37:17.915880 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 09:37:18 crc kubenswrapper[4715]: I1210 09:37:18.010308 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4e39b8d-4a24-486d-897c-39c46b733dec-kube-api-access\") pod \"installer-9-crc\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:18 crc kubenswrapper[4715]: I1210 09:37:18.010368 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:18 crc kubenswrapper[4715]: I1210 09:37:18.010408 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-var-lock\") pod \"installer-9-crc\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:18 crc kubenswrapper[4715]: I1210 09:37:18.418737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4e39b8d-4a24-486d-897c-39c46b733dec-kube-api-access\") pod \"installer-9-crc\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:18 crc kubenswrapper[4715]: I1210 09:37:18.418857 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:18 crc kubenswrapper[4715]: I1210 09:37:18.418911 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-var-lock\") pod \"installer-9-crc\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:18 crc kubenswrapper[4715]: I1210 09:37:18.419028 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:18 crc kubenswrapper[4715]: I1210 09:37:18.419175 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-var-lock\") pod \"installer-9-crc\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:18 crc kubenswrapper[4715]: I1210 09:37:18.446307 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4e39b8d-4a24-486d-897c-39c46b733dec-kube-api-access\") pod \"installer-9-crc\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:18 crc kubenswrapper[4715]: I1210 09:37:18.530194 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:37:18 crc kubenswrapper[4715]: E1210 09:37:18.788705 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 10 09:37:18 crc kubenswrapper[4715]: E1210 09:37:18.788961 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jgb4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-fx8wx_openshift-marketplace(078de70a-e498-487c-81a0-e7e3532e7c48): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 09:37:18 crc kubenswrapper[4715]: E1210 09:37:18.790139 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-fx8wx" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" Dec 10 09:37:19 crc kubenswrapper[4715]: I1210 09:37:19.506330 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5" exitCode=0 Dec 10 09:37:19 crc kubenswrapper[4715]: I1210 09:37:19.506374 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5"} Dec 10 09:37:19 crc kubenswrapper[4715]: E1210 09:37:19.641789 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 10 09:37:19 crc kubenswrapper[4715]: E1210 09:37:19.642322 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-btvn2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9n9gx_openshift-marketplace(efadbc66-5bb8-46c9-88cf-482defd875bd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 09:37:19 crc kubenswrapper[4715]: E1210 09:37:19.643579 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9n9gx" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" Dec 10 09:37:20 crc kubenswrapper[4715]: E1210 09:37:20.343259 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-fx8wx" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" Dec 10 09:37:20 crc kubenswrapper[4715]: E1210 09:37:20.418860 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 10 09:37:20 crc kubenswrapper[4715]: E1210 09:37:20.419028 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lbjkm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vq4cc_openshift-marketplace(fa81c629-c83d-444d-a113-de3681b8f0cc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 09:37:20 crc kubenswrapper[4715]: E1210 09:37:20.420348 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vq4cc" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" Dec 10 09:37:21 crc kubenswrapper[4715]: E1210 09:37:21.446875 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vq4cc" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" Dec 10 09:37:21 crc kubenswrapper[4715]: E1210 09:37:21.447678 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9n9gx" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" Dec 10 09:37:21 crc kubenswrapper[4715]: E1210 09:37:21.507725 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 10 09:37:21 crc kubenswrapper[4715]: E1210 09:37:21.508272 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p727r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8t5q2_openshift-marketplace(676753c5-d951-4722-a5ef-82e3227e0c7e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 09:37:21 crc kubenswrapper[4715]: E1210 09:37:21.510061 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-8t5q2" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" Dec 10 09:37:21 crc kubenswrapper[4715]: E1210 09:37:21.530926 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 10 09:37:21 crc kubenswrapper[4715]: E1210 09:37:21.531092 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-98xzg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xsbgs_openshift-marketplace(88de80a3-3467-4439-82ac-3a26dd5073b4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 09:37:21 crc kubenswrapper[4715]: E1210 09:37:21.534122 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xsbgs" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.306888 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xsbgs" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.306992 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8t5q2" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" Dec 10 09:37:24 crc kubenswrapper[4715]: I1210 09:37:24.369447 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:37:24 crc kubenswrapper[4715]: I1210 09:37:24.369507 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.397725 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.398107 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6zp86,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jrl25_openshift-marketplace(6d6da964-23be-49cc-8555-97edb12b5f28): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.397725 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.398284 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5t6ln,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6fxk6_openshift-marketplace(9083d000-4c4d-4e48-82d7-1f7a1ec885b0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.399500 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6fxk6" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.399500 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jrl25" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.422257 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.422408 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-42w9r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xn5c9_openshift-marketplace(8c078273-d5a5-4371-8ad5-091114f61311): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.424135 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xn5c9" podUID="8c078273-d5a5-4371-8ad5-091114f61311" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.547719 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jrl25" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.547855 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xn5c9" podUID="8c078273-d5a5-4371-8ad5-091114f61311" Dec 10 09:37:24 crc kubenswrapper[4715]: E1210 09:37:24.549679 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6fxk6" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" Dec 10 09:37:24 crc kubenswrapper[4715]: I1210 09:37:24.761458 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 09:37:24 crc kubenswrapper[4715]: I1210 09:37:24.862109 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 09:37:24 crc kubenswrapper[4715]: W1210 09:37:24.879518 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda4e39b8d_4a24_486d_897c_39c46b733dec.slice/crio-6cace447ac5abeef102610336a403257bfc049925fc91d9ade3e893d261ab743 WatchSource:0}: Error finding container 6cace447ac5abeef102610336a403257bfc049925fc91d9ade3e893d261ab743: Status 404 returned error can't find the container with id 6cace447ac5abeef102610336a403257bfc049925fc91d9ade3e893d261ab743 Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.554518 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b9k84" event={"ID":"55cadec6-a68b-4d03-abc0-d9c628c3e4d0","Type":"ContainerStarted","Data":"e4d13cf01e3893414e4019fa7b0aeea670139e8805faa35604c78d0acb5f6f76"} Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.555775 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-b9k84" Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.556523 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.556586 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.558073 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"f0e5a5883c2401a24c6955974c04dfd24314a74cbb02a997f9f8a68454c4153f"} Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.559644 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a4e39b8d-4a24-486d-897c-39c46b733dec","Type":"ContainerStarted","Data":"f2e36499884cd2b4008b5b906754b3c7e6a2471ae8d7ecc86271d1a366209fdd"} Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.559692 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a4e39b8d-4a24-486d-897c-39c46b733dec","Type":"ContainerStarted","Data":"6cace447ac5abeef102610336a403257bfc049925fc91d9ade3e893d261ab743"} Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.560930 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c19345d5-5575-4321-a37e-045244edec62","Type":"ContainerStarted","Data":"a90ca768c7db1e4a439522e8a94b4574972821b53ff425f415587e85746e5ddf"} Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.560954 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c19345d5-5575-4321-a37e-045244edec62","Type":"ContainerStarted","Data":"dae061699498cae8a6eb14eb32c3038d68f13a77e6731a35ef60a02304a99cea"} Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.610295 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=8.610272541 podStartE2EDuration="8.610272541s" podCreationTimestamp="2025-12-10 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:37:25.609811216 +0000 UTC m=+208.353357467" watchObservedRunningTime="2025-12-10 09:37:25.610272541 +0000 UTC m=+208.353818802" Dec 10 09:37:25 crc kubenswrapper[4715]: I1210 09:37:25.634240 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=13.634224433 podStartE2EDuration="13.634224433s" podCreationTimestamp="2025-12-10 09:37:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:37:25.633779149 +0000 UTC m=+208.377325400" watchObservedRunningTime="2025-12-10 09:37:25.634224433 +0000 UTC m=+208.377770674" Dec 10 09:37:26 crc kubenswrapper[4715]: I1210 09:37:26.568595 4715 generic.go:334] "Generic (PLEG): container finished" podID="c19345d5-5575-4321-a37e-045244edec62" containerID="a90ca768c7db1e4a439522e8a94b4574972821b53ff425f415587e85746e5ddf" exitCode=0 Dec 10 09:37:26 crc kubenswrapper[4715]: I1210 09:37:26.569064 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c19345d5-5575-4321-a37e-045244edec62","Type":"ContainerDied","Data":"a90ca768c7db1e4a439522e8a94b4574972821b53ff425f415587e85746e5ddf"} Dec 10 09:37:26 crc kubenswrapper[4715]: I1210 09:37:26.569490 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-b9k84 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 10 09:37:26 crc kubenswrapper[4715]: I1210 09:37:26.569519 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b9k84" podUID="55cadec6-a68b-4d03-abc0-d9c628c3e4d0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 10 09:37:28 crc kubenswrapper[4715]: I1210 09:37:28.158074 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 09:37:28 crc kubenswrapper[4715]: I1210 09:37:28.257276 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c19345d5-5575-4321-a37e-045244edec62-kube-api-access\") pod \"c19345d5-5575-4321-a37e-045244edec62\" (UID: \"c19345d5-5575-4321-a37e-045244edec62\") " Dec 10 09:37:28 crc kubenswrapper[4715]: I1210 09:37:28.257405 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c19345d5-5575-4321-a37e-045244edec62-kubelet-dir\") pod \"c19345d5-5575-4321-a37e-045244edec62\" (UID: \"c19345d5-5575-4321-a37e-045244edec62\") " Dec 10 09:37:28 crc kubenswrapper[4715]: I1210 09:37:28.257690 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c19345d5-5575-4321-a37e-045244edec62-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c19345d5-5575-4321-a37e-045244edec62" (UID: "c19345d5-5575-4321-a37e-045244edec62"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:37:28 crc kubenswrapper[4715]: I1210 09:37:28.359075 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c19345d5-5575-4321-a37e-045244edec62-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:37:28 crc kubenswrapper[4715]: I1210 09:37:28.379074 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19345d5-5575-4321-a37e-045244edec62-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c19345d5-5575-4321-a37e-045244edec62" (UID: "c19345d5-5575-4321-a37e-045244edec62"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:37:28 crc kubenswrapper[4715]: I1210 09:37:28.459442 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c19345d5-5575-4321-a37e-045244edec62-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 09:37:28 crc kubenswrapper[4715]: I1210 09:37:28.582962 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c19345d5-5575-4321-a37e-045244edec62","Type":"ContainerDied","Data":"dae061699498cae8a6eb14eb32c3038d68f13a77e6731a35ef60a02304a99cea"} Dec 10 09:37:28 crc kubenswrapper[4715]: I1210 09:37:28.583011 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dae061699498cae8a6eb14eb32c3038d68f13a77e6731a35ef60a02304a99cea" Dec 10 09:37:28 crc kubenswrapper[4715]: I1210 09:37:28.583100 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 09:37:34 crc kubenswrapper[4715]: I1210 09:37:34.375593 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-b9k84" Dec 10 09:37:47 crc kubenswrapper[4715]: I1210 09:37:47.359194 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrl25" event={"ID":"6d6da964-23be-49cc-8555-97edb12b5f28","Type":"ContainerStarted","Data":"c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7"} Dec 10 09:37:47 crc kubenswrapper[4715]: I1210 09:37:47.363302 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq4cc" event={"ID":"fa81c629-c83d-444d-a113-de3681b8f0cc","Type":"ContainerStarted","Data":"469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87"} Dec 10 09:37:47 crc kubenswrapper[4715]: I1210 09:37:47.369099 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fx8wx" event={"ID":"078de70a-e498-487c-81a0-e7e3532e7c48","Type":"ContainerStarted","Data":"80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e"} Dec 10 09:37:47 crc kubenswrapper[4715]: I1210 09:37:47.372283 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9n9gx" event={"ID":"efadbc66-5bb8-46c9-88cf-482defd875bd","Type":"ContainerStarted","Data":"4760e97846cac74e6098339743042376aa198fef06acf2621095213fa30a4c97"} Dec 10 09:37:47 crc kubenswrapper[4715]: I1210 09:37:47.381755 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fxk6" event={"ID":"9083d000-4c4d-4e48-82d7-1f7a1ec885b0","Type":"ContainerStarted","Data":"d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a"} Dec 10 09:37:47 crc kubenswrapper[4715]: I1210 09:37:47.388897 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xn5c9" event={"ID":"8c078273-d5a5-4371-8ad5-091114f61311","Type":"ContainerStarted","Data":"f11b5d4926fe1edf270cd26217e42e0d7308218583d6e3bdfeccbcd35be5ff4f"} Dec 10 09:37:47 crc kubenswrapper[4715]: I1210 09:37:47.402596 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t5q2" event={"ID":"676753c5-d951-4722-a5ef-82e3227e0c7e","Type":"ContainerStarted","Data":"e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc"} Dec 10 09:37:48 crc kubenswrapper[4715]: I1210 09:37:48.415744 4715 generic.go:334] "Generic (PLEG): container finished" podID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerID="58ad0665f963c1137cdd67194611d725eabdb3eba9ba045d92f552eb9dda471b" exitCode=0 Dec 10 09:37:48 crc kubenswrapper[4715]: I1210 09:37:48.416052 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsbgs" event={"ID":"88de80a3-3467-4439-82ac-3a26dd5073b4","Type":"ContainerDied","Data":"58ad0665f963c1137cdd67194611d725eabdb3eba9ba045d92f552eb9dda471b"} Dec 10 09:37:48 crc kubenswrapper[4715]: I1210 09:37:48.426449 4715 generic.go:334] "Generic (PLEG): container finished" podID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerID="e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc" exitCode=0 Dec 10 09:37:48 crc kubenswrapper[4715]: I1210 09:37:48.426500 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t5q2" event={"ID":"676753c5-d951-4722-a5ef-82e3227e0c7e","Type":"ContainerDied","Data":"e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc"} Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.499121 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsbgs" event={"ID":"88de80a3-3467-4439-82ac-3a26dd5073b4","Type":"ContainerStarted","Data":"a5528b69146097448e4b71a784a661f5ff1c3d74d953520c183876f7d2f842c5"} Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.502661 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t5q2" event={"ID":"676753c5-d951-4722-a5ef-82e3227e0c7e","Type":"ContainerStarted","Data":"b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64"} Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.505256 4715 generic.go:334] "Generic (PLEG): container finished" podID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerID="469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87" exitCode=0 Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.506305 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq4cc" event={"ID":"fa81c629-c83d-444d-a113-de3681b8f0cc","Type":"ContainerDied","Data":"469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87"} Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.513371 4715 generic.go:334] "Generic (PLEG): container finished" podID="078de70a-e498-487c-81a0-e7e3532e7c48" containerID="80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e" exitCode=0 Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.513465 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fx8wx" event={"ID":"078de70a-e498-487c-81a0-e7e3532e7c48","Type":"ContainerDied","Data":"80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e"} Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.516221 4715 generic.go:334] "Generic (PLEG): container finished" podID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerID="4760e97846cac74e6098339743042376aa198fef06acf2621095213fa30a4c97" exitCode=0 Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.516290 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9n9gx" event={"ID":"efadbc66-5bb8-46c9-88cf-482defd875bd","Type":"ContainerDied","Data":"4760e97846cac74e6098339743042376aa198fef06acf2621095213fa30a4c97"} Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.519157 4715 generic.go:334] "Generic (PLEG): container finished" podID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerID="d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a" exitCode=0 Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.519200 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fxk6" event={"ID":"9083d000-4c4d-4e48-82d7-1f7a1ec885b0","Type":"ContainerDied","Data":"d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a"} Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.548163 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xsbgs" podStartSLOduration=3.161223102 podStartE2EDuration="1m13.548143256s" podCreationTimestamp="2025-12-10 09:36:36 +0000 UTC" firstStartedPulling="2025-12-10 09:36:38.584583502 +0000 UTC m=+161.328129753" lastFinishedPulling="2025-12-10 09:37:48.971503656 +0000 UTC m=+231.715049907" observedRunningTime="2025-12-10 09:37:49.54764558 +0000 UTC m=+232.291191841" watchObservedRunningTime="2025-12-10 09:37:49.548143256 +0000 UTC m=+232.291689507" Dec 10 09:37:49 crc kubenswrapper[4715]: I1210 09:37:49.666629 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8t5q2" podStartSLOduration=3.218749768 podStartE2EDuration="1m13.666608327s" podCreationTimestamp="2025-12-10 09:36:36 +0000 UTC" firstStartedPulling="2025-12-10 09:36:38.593806142 +0000 UTC m=+161.337352393" lastFinishedPulling="2025-12-10 09:37:49.041664701 +0000 UTC m=+231.785210952" observedRunningTime="2025-12-10 09:37:49.661451407 +0000 UTC m=+232.404997658" watchObservedRunningTime="2025-12-10 09:37:49.666608327 +0000 UTC m=+232.410154578" Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.528145 4715 generic.go:334] "Generic (PLEG): container finished" podID="6d6da964-23be-49cc-8555-97edb12b5f28" containerID="c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7" exitCode=0 Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.528232 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrl25" event={"ID":"6d6da964-23be-49cc-8555-97edb12b5f28","Type":"ContainerDied","Data":"c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7"} Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.532761 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq4cc" event={"ID":"fa81c629-c83d-444d-a113-de3681b8f0cc","Type":"ContainerStarted","Data":"28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40"} Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.538064 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fx8wx" event={"ID":"078de70a-e498-487c-81a0-e7e3532e7c48","Type":"ContainerStarted","Data":"cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0"} Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.540474 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9n9gx" event={"ID":"efadbc66-5bb8-46c9-88cf-482defd875bd","Type":"ContainerStarted","Data":"13e537ef20ecf4ca4643eb32c23d2e08592fa9082cc3f151c2887ff8b93e4150"} Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.542763 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fxk6" event={"ID":"9083d000-4c4d-4e48-82d7-1f7a1ec885b0","Type":"ContainerStarted","Data":"c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50"} Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.545675 4715 generic.go:334] "Generic (PLEG): container finished" podID="8c078273-d5a5-4371-8ad5-091114f61311" containerID="f11b5d4926fe1edf270cd26217e42e0d7308218583d6e3bdfeccbcd35be5ff4f" exitCode=0 Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.545711 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xn5c9" event={"ID":"8c078273-d5a5-4371-8ad5-091114f61311","Type":"ContainerDied","Data":"f11b5d4926fe1edf270cd26217e42e0d7308218583d6e3bdfeccbcd35be5ff4f"} Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.572568 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fx8wx" podStartSLOduration=4.128186825 podStartE2EDuration="1m16.57255219s" podCreationTimestamp="2025-12-10 09:36:34 +0000 UTC" firstStartedPulling="2025-12-10 09:36:37.575391576 +0000 UTC m=+160.318937827" lastFinishedPulling="2025-12-10 09:37:50.019756941 +0000 UTC m=+232.763303192" observedRunningTime="2025-12-10 09:37:50.569575878 +0000 UTC m=+233.313122129" watchObservedRunningTime="2025-12-10 09:37:50.57255219 +0000 UTC m=+233.316098441" Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.595359 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9n9gx" podStartSLOduration=5.093735124 podStartE2EDuration="1m16.595339487s" podCreationTimestamp="2025-12-10 09:36:34 +0000 UTC" firstStartedPulling="2025-12-10 09:36:38.576323231 +0000 UTC m=+161.319869482" lastFinishedPulling="2025-12-10 09:37:50.077927594 +0000 UTC m=+232.821473845" observedRunningTime="2025-12-10 09:37:50.590783945 +0000 UTC m=+233.334330216" watchObservedRunningTime="2025-12-10 09:37:50.595339487 +0000 UTC m=+233.338885738" Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.616969 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vq4cc" podStartSLOduration=5.295857502 podStartE2EDuration="1m16.616950176s" podCreationTimestamp="2025-12-10 09:36:34 +0000 UTC" firstStartedPulling="2025-12-10 09:36:38.618310377 +0000 UTC m=+161.361856638" lastFinishedPulling="2025-12-10 09:37:49.939403061 +0000 UTC m=+232.682949312" observedRunningTime="2025-12-10 09:37:50.612761437 +0000 UTC m=+233.356307698" watchObservedRunningTime="2025-12-10 09:37:50.616950176 +0000 UTC m=+233.360496437" Dec 10 09:37:50 crc kubenswrapper[4715]: I1210 09:37:50.658650 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6fxk6" podStartSLOduration=5.011342154 podStartE2EDuration="1m16.658632888s" podCreationTimestamp="2025-12-10 09:36:34 +0000 UTC" firstStartedPulling="2025-12-10 09:36:38.581335538 +0000 UTC m=+161.324881779" lastFinishedPulling="2025-12-10 09:37:50.228626262 +0000 UTC m=+232.972172513" observedRunningTime="2025-12-10 09:37:50.657241065 +0000 UTC m=+233.400787316" watchObservedRunningTime="2025-12-10 09:37:50.658632888 +0000 UTC m=+233.402179139" Dec 10 09:37:51 crc kubenswrapper[4715]: I1210 09:37:51.554665 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrl25" event={"ID":"6d6da964-23be-49cc-8555-97edb12b5f28","Type":"ContainerStarted","Data":"2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea"} Dec 10 09:37:51 crc kubenswrapper[4715]: I1210 09:37:51.574367 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jrl25" podStartSLOduration=2.240434696 podStartE2EDuration="1m14.574347265s" podCreationTimestamp="2025-12-10 09:36:37 +0000 UTC" firstStartedPulling="2025-12-10 09:36:38.612520258 +0000 UTC m=+161.356066499" lastFinishedPulling="2025-12-10 09:37:50.946432817 +0000 UTC m=+233.689979068" observedRunningTime="2025-12-10 09:37:51.572871609 +0000 UTC m=+234.316417870" watchObservedRunningTime="2025-12-10 09:37:51.574347265 +0000 UTC m=+234.317893516" Dec 10 09:37:52 crc kubenswrapper[4715]: I1210 09:37:52.562490 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xn5c9" event={"ID":"8c078273-d5a5-4371-8ad5-091114f61311","Type":"ContainerStarted","Data":"4eb9a31a1ec94d641cb6626ecc3ac12dcd709fdb2e8d10b8c871f9e31e25e010"} Dec 10 09:37:52 crc kubenswrapper[4715]: I1210 09:37:52.582011 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xn5c9" podStartSLOduration=3.216663851 podStartE2EDuration="1m15.58195976s" podCreationTimestamp="2025-12-10 09:36:37 +0000 UTC" firstStartedPulling="2025-12-10 09:36:39.711689982 +0000 UTC m=+162.455236233" lastFinishedPulling="2025-12-10 09:37:52.076985891 +0000 UTC m=+234.820532142" observedRunningTime="2025-12-10 09:37:52.579539435 +0000 UTC m=+235.323085696" watchObservedRunningTime="2025-12-10 09:37:52.58195976 +0000 UTC m=+235.325506021" Dec 10 09:37:54 crc kubenswrapper[4715]: I1210 09:37:54.758679 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:37:54 crc kubenswrapper[4715]: I1210 09:37:54.760122 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:37:54 crc kubenswrapper[4715]: I1210 09:37:54.884388 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:37:54 crc kubenswrapper[4715]: I1210 09:37:54.884450 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:37:55 crc kubenswrapper[4715]: I1210 09:37:55.000509 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:37:55 crc kubenswrapper[4715]: I1210 09:37:55.001242 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:37:55 crc kubenswrapper[4715]: I1210 09:37:55.163981 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:37:55 crc kubenswrapper[4715]: I1210 09:37:55.164053 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:37:56 crc kubenswrapper[4715]: I1210 09:37:56.758046 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:37:56 crc kubenswrapper[4715]: I1210 09:37:56.758115 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.149439 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.149876 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.150053 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.161640 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.166206 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.167337 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.167373 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.216387 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.217050 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.221682 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.223564 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.247535 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.630674 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.665809 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.756251 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.756603 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.801762 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6fxk6"] Dec 10 09:37:57 crc kubenswrapper[4715]: I1210 09:37:57.804672 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:37:58 crc kubenswrapper[4715]: I1210 09:37:58.110413 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:37:58 crc kubenswrapper[4715]: I1210 09:37:58.110474 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:37:58 crc kubenswrapper[4715]: I1210 09:37:58.159873 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:37:58 crc kubenswrapper[4715]: I1210 09:37:58.595253 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6fxk6" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerName="registry-server" containerID="cri-o://c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50" gracePeriod=2 Dec 10 09:37:58 crc kubenswrapper[4715]: I1210 09:37:58.642638 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:37:58 crc kubenswrapper[4715]: I1210 09:37:58.650573 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.397338 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.512317 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t6ln\" (UniqueName: \"kubernetes.io/projected/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-kube-api-access-5t6ln\") pod \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.512418 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-catalog-content\") pod \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.514197 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-utilities" (OuterVolumeSpecName: "utilities") pod "9083d000-4c4d-4e48-82d7-1f7a1ec885b0" (UID: "9083d000-4c4d-4e48-82d7-1f7a1ec885b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.515062 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-utilities\") pod \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\" (UID: \"9083d000-4c4d-4e48-82d7-1f7a1ec885b0\") " Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.515774 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.520136 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-kube-api-access-5t6ln" (OuterVolumeSpecName: "kube-api-access-5t6ln") pod "9083d000-4c4d-4e48-82d7-1f7a1ec885b0" (UID: "9083d000-4c4d-4e48-82d7-1f7a1ec885b0"). InnerVolumeSpecName "kube-api-access-5t6ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.598745 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9n9gx"] Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.599365 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9n9gx" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerName="registry-server" containerID="cri-o://13e537ef20ecf4ca4643eb32c23d2e08592fa9082cc3f151c2887ff8b93e4150" gracePeriod=2 Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.605640 4715 generic.go:334] "Generic (PLEG): container finished" podID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerID="c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50" exitCode=0 Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.606161 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fxk6" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.616777 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t6ln\" (UniqueName: \"kubernetes.io/projected/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-kube-api-access-5t6ln\") on node \"crc\" DevicePath \"\"" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.616961 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fxk6" event={"ID":"9083d000-4c4d-4e48-82d7-1f7a1ec885b0","Type":"ContainerDied","Data":"c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50"} Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.617035 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fxk6" event={"ID":"9083d000-4c4d-4e48-82d7-1f7a1ec885b0","Type":"ContainerDied","Data":"cf4973b416d6f947364eb825c31b4e6dd0a5ca17b165b4ea0f9dbfd420b3f052"} Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.617059 4715 scope.go:117] "RemoveContainer" containerID="c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.785388 4715 scope.go:117] "RemoveContainer" containerID="d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.804898 4715 scope.go:117] "RemoveContainer" containerID="babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.825260 4715 scope.go:117] "RemoveContainer" containerID="c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50" Dec 10 09:37:59 crc kubenswrapper[4715]: E1210 09:37:59.825659 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50\": container with ID starting with c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50 not found: ID does not exist" containerID="c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.825702 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50"} err="failed to get container status \"c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50\": rpc error: code = NotFound desc = could not find container \"c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50\": container with ID starting with c387f7a1d5e68c14b3523fdea492450d0fb9204a741c1613a7014da60ba52d50 not found: ID does not exist" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.825730 4715 scope.go:117] "RemoveContainer" containerID="d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a" Dec 10 09:37:59 crc kubenswrapper[4715]: E1210 09:37:59.826047 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a\": container with ID starting with d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a not found: ID does not exist" containerID="d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.826084 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a"} err="failed to get container status \"d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a\": rpc error: code = NotFound desc = could not find container \"d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a\": container with ID starting with d8729129e541828f80fc34505476b03cbb409fd5373043259605895eede9713a not found: ID does not exist" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.826110 4715 scope.go:117] "RemoveContainer" containerID="babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9" Dec 10 09:37:59 crc kubenswrapper[4715]: E1210 09:37:59.826388 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9\": container with ID starting with babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9 not found: ID does not exist" containerID="babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.826424 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9"} err="failed to get container status \"babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9\": rpc error: code = NotFound desc = could not find container \"babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9\": container with ID starting with babc62cfb1014508bfc9eecb316e9e3c57d083e54a8aadb57c94f8ff2cf498f9 not found: ID does not exist" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.839330 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9083d000-4c4d-4e48-82d7-1f7a1ec885b0" (UID: "9083d000-4c4d-4e48-82d7-1f7a1ec885b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.921102 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9083d000-4c4d-4e48-82d7-1f7a1ec885b0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.937531 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6fxk6"] Dec 10 09:37:59 crc kubenswrapper[4715]: I1210 09:37:59.941466 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6fxk6"] Dec 10 09:37:59 crc kubenswrapper[4715]: E1210 09:37:59.984392 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9083d000_4c4d_4e48_82d7_1f7a1ec885b0.slice/crio-cf4973b416d6f947364eb825c31b4e6dd0a5ca17b165b4ea0f9dbfd420b3f052\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9083d000_4c4d_4e48_82d7_1f7a1ec885b0.slice\": RecentStats: unable to find data in memory cache]" Dec 10 09:38:00 crc kubenswrapper[4715]: I1210 09:38:00.198761 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsbgs"] Dec 10 09:38:00 crc kubenswrapper[4715]: I1210 09:38:00.198992 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xsbgs" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerName="registry-server" containerID="cri-o://a5528b69146097448e4b71a784a661f5ff1c3d74d953520c183876f7d2f842c5" gracePeriod=2 Dec 10 09:38:00 crc kubenswrapper[4715]: I1210 09:38:00.614488 4715 generic.go:334] "Generic (PLEG): container finished" podID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerID="13e537ef20ecf4ca4643eb32c23d2e08592fa9082cc3f151c2887ff8b93e4150" exitCode=0 Dec 10 09:38:00 crc kubenswrapper[4715]: I1210 09:38:00.614571 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9n9gx" event={"ID":"efadbc66-5bb8-46c9-88cf-482defd875bd","Type":"ContainerDied","Data":"13e537ef20ecf4ca4643eb32c23d2e08592fa9082cc3f151c2887ff8b93e4150"} Dec 10 09:38:00 crc kubenswrapper[4715]: I1210 09:38:00.618617 4715 generic.go:334] "Generic (PLEG): container finished" podID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerID="a5528b69146097448e4b71a784a661f5ff1c3d74d953520c183876f7d2f842c5" exitCode=0 Dec 10 09:38:00 crc kubenswrapper[4715]: I1210 09:38:00.618711 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsbgs" event={"ID":"88de80a3-3467-4439-82ac-3a26dd5073b4","Type":"ContainerDied","Data":"a5528b69146097448e4b71a784a661f5ff1c3d74d953520c183876f7d2f842c5"} Dec 10 09:38:00 crc kubenswrapper[4715]: I1210 09:38:00.847522 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.037425 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-utilities\") pod \"efadbc66-5bb8-46c9-88cf-482defd875bd\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.037937 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btvn2\" (UniqueName: \"kubernetes.io/projected/efadbc66-5bb8-46c9-88cf-482defd875bd-kube-api-access-btvn2\") pod \"efadbc66-5bb8-46c9-88cf-482defd875bd\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.037987 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-catalog-content\") pod \"efadbc66-5bb8-46c9-88cf-482defd875bd\" (UID: \"efadbc66-5bb8-46c9-88cf-482defd875bd\") " Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.039210 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-utilities" (OuterVolumeSpecName: "utilities") pod "efadbc66-5bb8-46c9-88cf-482defd875bd" (UID: "efadbc66-5bb8-46c9-88cf-482defd875bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.041174 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.049844 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efadbc66-5bb8-46c9-88cf-482defd875bd-kube-api-access-btvn2" (OuterVolumeSpecName: "kube-api-access-btvn2") pod "efadbc66-5bb8-46c9-88cf-482defd875bd" (UID: "efadbc66-5bb8-46c9-88cf-482defd875bd"). InnerVolumeSpecName "kube-api-access-btvn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.104125 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "efadbc66-5bb8-46c9-88cf-482defd875bd" (UID: "efadbc66-5bb8-46c9-88cf-482defd875bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.139437 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98xzg\" (UniqueName: \"kubernetes.io/projected/88de80a3-3467-4439-82ac-3a26dd5073b4-kube-api-access-98xzg\") pod \"88de80a3-3467-4439-82ac-3a26dd5073b4\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.139828 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-utilities\") pod \"88de80a3-3467-4439-82ac-3a26dd5073b4\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.139958 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-catalog-content\") pod \"88de80a3-3467-4439-82ac-3a26dd5073b4\" (UID: \"88de80a3-3467-4439-82ac-3a26dd5073b4\") " Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.140238 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btvn2\" (UniqueName: \"kubernetes.io/projected/efadbc66-5bb8-46c9-88cf-482defd875bd-kube-api-access-btvn2\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.140367 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.140474 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efadbc66-5bb8-46c9-88cf-482defd875bd-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.140412 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-utilities" (OuterVolumeSpecName: "utilities") pod "88de80a3-3467-4439-82ac-3a26dd5073b4" (UID: "88de80a3-3467-4439-82ac-3a26dd5073b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.142114 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88de80a3-3467-4439-82ac-3a26dd5073b4-kube-api-access-98xzg" (OuterVolumeSpecName: "kube-api-access-98xzg") pod "88de80a3-3467-4439-82ac-3a26dd5073b4" (UID: "88de80a3-3467-4439-82ac-3a26dd5073b4"). InnerVolumeSpecName "kube-api-access-98xzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.158969 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88de80a3-3467-4439-82ac-3a26dd5073b4" (UID: "88de80a3-3467-4439-82ac-3a26dd5073b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.241600 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98xzg\" (UniqueName: \"kubernetes.io/projected/88de80a3-3467-4439-82ac-3a26dd5073b4-kube-api-access-98xzg\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.241674 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.241688 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88de80a3-3467-4439-82ac-3a26dd5073b4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.612610 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" path="/var/lib/kubelet/pods/9083d000-4c4d-4e48-82d7-1f7a1ec885b0/volumes" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.624813 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9n9gx" event={"ID":"efadbc66-5bb8-46c9-88cf-482defd875bd","Type":"ContainerDied","Data":"bc0d4f8c240d9c68ca7b68ce9709e02f5c3bbaac9192e93c89679ccdc74e40d0"} Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.624860 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9n9gx" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.624874 4715 scope.go:117] "RemoveContainer" containerID="13e537ef20ecf4ca4643eb32c23d2e08592fa9082cc3f151c2887ff8b93e4150" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.627779 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xsbgs" event={"ID":"88de80a3-3467-4439-82ac-3a26dd5073b4","Type":"ContainerDied","Data":"ccfbdaf22d9428c76a62c87d47bc1278ceb0f07881eed905270f7873f7f6cdec"} Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.628127 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xsbgs" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.653599 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsbgs"] Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.656508 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xsbgs"] Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.660983 4715 scope.go:117] "RemoveContainer" containerID="4760e97846cac74e6098339743042376aa198fef06acf2621095213fa30a4c97" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.662719 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9n9gx"] Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.665345 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9n9gx"] Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.681092 4715 scope.go:117] "RemoveContainer" containerID="23b1f5fe5e829f796334cefb6d23e35fe4e33c7095fdd0759ac4f3ebb24047d0" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.697274 4715 scope.go:117] "RemoveContainer" containerID="a5528b69146097448e4b71a784a661f5ff1c3d74d953520c183876f7d2f842c5" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.712765 4715 scope.go:117] "RemoveContainer" containerID="58ad0665f963c1137cdd67194611d725eabdb3eba9ba045d92f552eb9dda471b" Dec 10 09:38:01 crc kubenswrapper[4715]: I1210 09:38:01.726590 4715 scope.go:117] "RemoveContainer" containerID="37f48a809b5c9f81628cd8ddfb6265ee7b5016a9f614653652482092bd2a0976" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.600308 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xn5c9"] Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.600532 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xn5c9" podUID="8c078273-d5a5-4371-8ad5-091114f61311" containerName="registry-server" containerID="cri-o://4eb9a31a1ec94d641cb6626ecc3ac12dcd709fdb2e8d10b8c871f9e31e25e010" gracePeriod=2 Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.870807 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.871048 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerName="extract-utilities" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871060 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerName="extract-utilities" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.871070 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerName="extract-content" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871076 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerName="extract-content" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.871087 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerName="extract-content" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871092 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerName="extract-content" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.871101 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerName="extract-utilities" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871106 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerName="extract-utilities" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.871116 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerName="registry-server" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871122 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerName="registry-server" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.871132 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerName="registry-server" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871137 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerName="registry-server" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.871144 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerName="registry-server" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871149 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerName="registry-server" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.871156 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerName="extract-utilities" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871161 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerName="extract-utilities" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.871167 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19345d5-5575-4321-a37e-045244edec62" containerName="pruner" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871173 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19345d5-5575-4321-a37e-045244edec62" containerName="pruner" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.871182 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerName="extract-content" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871188 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerName="extract-content" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871275 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19345d5-5575-4321-a37e-045244edec62" containerName="pruner" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871284 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" containerName="registry-server" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871292 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" containerName="registry-server" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871298 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9083d000-4c4d-4e48-82d7-1f7a1ec885b0" containerName="registry-server" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.871662 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.872479 4715 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.872794 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed" gracePeriod=15 Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.872866 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335" gracePeriod=15 Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.872953 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d" gracePeriod=15 Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.872786 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee" gracePeriod=15 Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.872941 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338" gracePeriod=15 Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.873744 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.874240 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874263 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.874276 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874284 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.874313 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874321 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.874336 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874344 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.874357 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874364 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.874399 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874408 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 09:38:02 crc kubenswrapper[4715]: E1210 09:38:02.874418 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874426 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874557 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874577 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874587 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874599 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874608 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.874831 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.912014 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.963659 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.963690 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.963716 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.963749 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:02 crc kubenswrapper[4715]: I1210 09:38:02.963815 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065167 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065222 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065243 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065260 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065278 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065308 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065331 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065353 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065464 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065498 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065520 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065629 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.065649 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: E1210 09:38:03.144428 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:03 crc kubenswrapper[4715]: E1210 09:38:03.145062 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:03 crc kubenswrapper[4715]: E1210 09:38:03.145642 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:03 crc kubenswrapper[4715]: E1210 09:38:03.145996 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:03 crc kubenswrapper[4715]: E1210 09:38:03.146310 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.146349 4715 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 10 09:38:03 crc kubenswrapper[4715]: E1210 09:38:03.146639 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="200ms" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.166379 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.166438 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.166497 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.166535 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.166590 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.166622 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.203347 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:03 crc kubenswrapper[4715]: W1210 09:38:03.219217 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-139816b1ddc75d0d5bcc7a833b4402588f774ca31c05397141f448cb01cc43bd WatchSource:0}: Error finding container 139816b1ddc75d0d5bcc7a833b4402588f774ca31c05397141f448cb01cc43bd: Status 404 returned error can't find the container with id 139816b1ddc75d0d5bcc7a833b4402588f774ca31c05397141f448cb01cc43bd Dec 10 09:38:03 crc kubenswrapper[4715]: E1210 09:38:03.222079 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.201:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187fd118c2403aee openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 09:38:03.220933358 +0000 UTC m=+245.964479619,LastTimestamp:2025-12-10 09:38:03.220933358 +0000 UTC m=+245.964479619,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 09:38:03 crc kubenswrapper[4715]: E1210 09:38:03.347816 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="400ms" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.611660 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88de80a3-3467-4439-82ac-3a26dd5073b4" path="/var/lib/kubelet/pods/88de80a3-3467-4439-82ac-3a26dd5073b4/volumes" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.612839 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efadbc66-5bb8-46c9-88cf-482defd875bd" path="/var/lib/kubelet/pods/efadbc66-5bb8-46c9-88cf-482defd875bd/volumes" Dec 10 09:38:03 crc kubenswrapper[4715]: I1210 09:38:03.646867 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"139816b1ddc75d0d5bcc7a833b4402588f774ca31c05397141f448cb01cc43bd"} Dec 10 09:38:03 crc kubenswrapper[4715]: E1210 09:38:03.749248 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="800ms" Dec 10 09:38:04 crc kubenswrapper[4715]: E1210 09:38:04.021340 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:38:04Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:38:04Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:38:04Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:38:04Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:324d8852a972c066be00776701f7ea5f0b533e79431c220f6f6eb0b6b338ace7\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:cb4d03c3e9f89cd7fa16144ef8515fd5e1f3a5505d707667dea76b4e3cc09f9b\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1626095435},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:040ed5bd62bba86f603d3f5f7e1d29d3fab894f04c561f0d97c907340285e79e\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:99471ee5d5867f4757df6da943e7254536ab2a616bf2c62dc6900f0e1ea07109\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1215778122},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:26ea35413bef0c078547a03dc093f1d4f15a7d9fc91f05c687b1a437352f0856\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:ba6f8ed6f9b58e63d979cc9729f5ce2c6f22e99b35ce91b902e198eb78d6b106\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201960779},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1b1026c62413fa239fa4ff6541fe8bda656c1281867ad6ee2c848feccb13c97e\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:2b633ebdc901d19290af4dc2d09e2b59c504c0fc15a3fba410b0ce098e2d5753\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1141987142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:04 crc kubenswrapper[4715]: E1210 09:38:04.021962 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:04 crc kubenswrapper[4715]: E1210 09:38:04.022524 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:04 crc kubenswrapper[4715]: E1210 09:38:04.022892 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:04 crc kubenswrapper[4715]: E1210 09:38:04.023294 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:04 crc kubenswrapper[4715]: E1210 09:38:04.023326 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:38:04 crc kubenswrapper[4715]: E1210 09:38:04.550816 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="1.6s" Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.652953 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xn5c9" event={"ID":"8c078273-d5a5-4371-8ad5-091114f61311","Type":"ContainerDied","Data":"4eb9a31a1ec94d641cb6626ecc3ac12dcd709fdb2e8d10b8c871f9e31e25e010"} Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.652982 4715 generic.go:334] "Generic (PLEG): container finished" podID="8c078273-d5a5-4371-8ad5-091114f61311" containerID="4eb9a31a1ec94d641cb6626ecc3ac12dcd709fdb2e8d10b8c871f9e31e25e010" exitCode=0 Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.655249 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.656464 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.657168 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed" exitCode=0 Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.657193 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335" exitCode=0 Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.657203 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d" exitCode=0 Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.657213 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338" exitCode=2 Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.657250 4715 scope.go:117] "RemoveContainer" containerID="6e481dfe68614298f96612420759fe28a6803c69f3b715020139428fceccec47" Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.658672 4715 generic.go:334] "Generic (PLEG): container finished" podID="a4e39b8d-4a24-486d-897c-39c46b733dec" containerID="f2e36499884cd2b4008b5b906754b3c7e6a2471ae8d7ecc86271d1a366209fdd" exitCode=0 Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.658728 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a4e39b8d-4a24-486d-897c-39c46b733dec","Type":"ContainerDied","Data":"f2e36499884cd2b4008b5b906754b3c7e6a2471ae8d7ecc86271d1a366209fdd"} Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.659476 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.659855 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.663539 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204"} Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.664397 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:04 crc kubenswrapper[4715]: I1210 09:38:04.664723 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.903525 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.904982 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.905375 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.907029 4715 status_manager.go:851] "Failed to get status for pod" podUID="8c078273-d5a5-4371-8ad5-091114f61311" pod="openshift-marketplace/redhat-operators-xn5c9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xn5c9\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.947064 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.947643 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.948086 4715 status_manager.go:851] "Failed to get status for pod" podUID="8c078273-d5a5-4371-8ad5-091114f61311" pod="openshift-marketplace/redhat-operators-xn5c9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xn5c9\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.948361 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.999681 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42w9r\" (UniqueName: \"kubernetes.io/projected/8c078273-d5a5-4371-8ad5-091114f61311-kube-api-access-42w9r\") pod \"8c078273-d5a5-4371-8ad5-091114f61311\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.999742 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-catalog-content\") pod \"8c078273-d5a5-4371-8ad5-091114f61311\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " Dec 10 09:38:05 crc kubenswrapper[4715]: I1210 09:38:05.999790 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-utilities\") pod \"8c078273-d5a5-4371-8ad5-091114f61311\" (UID: \"8c078273-d5a5-4371-8ad5-091114f61311\") " Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.001479 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-utilities" (OuterVolumeSpecName: "utilities") pod "8c078273-d5a5-4371-8ad5-091114f61311" (UID: "8c078273-d5a5-4371-8ad5-091114f61311"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.006538 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c078273-d5a5-4371-8ad5-091114f61311-kube-api-access-42w9r" (OuterVolumeSpecName: "kube-api-access-42w9r") pod "8c078273-d5a5-4371-8ad5-091114f61311" (UID: "8c078273-d5a5-4371-8ad5-091114f61311"). InnerVolumeSpecName "kube-api-access-42w9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.100491 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4e39b8d-4a24-486d-897c-39c46b733dec-kube-api-access\") pod \"a4e39b8d-4a24-486d-897c-39c46b733dec\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.100567 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-kubelet-dir\") pod \"a4e39b8d-4a24-486d-897c-39c46b733dec\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.100593 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-var-lock\") pod \"a4e39b8d-4a24-486d-897c-39c46b733dec\" (UID: \"a4e39b8d-4a24-486d-897c-39c46b733dec\") " Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.100850 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42w9r\" (UniqueName: \"kubernetes.io/projected/8c078273-d5a5-4371-8ad5-091114f61311-kube-api-access-42w9r\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.100869 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.100906 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-var-lock" (OuterVolumeSpecName: "var-lock") pod "a4e39b8d-4a24-486d-897c-39c46b733dec" (UID: "a4e39b8d-4a24-486d-897c-39c46b733dec"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.100907 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a4e39b8d-4a24-486d-897c-39c46b733dec" (UID: "a4e39b8d-4a24-486d-897c-39c46b733dec"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.104543 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4e39b8d-4a24-486d-897c-39c46b733dec-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a4e39b8d-4a24-486d-897c-39c46b733dec" (UID: "a4e39b8d-4a24-486d-897c-39c46b733dec"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:38:06 crc kubenswrapper[4715]: E1210 09:38:06.152321 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="3.2s" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.202233 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4e39b8d-4a24-486d-897c-39c46b733dec-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.202270 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.202279 4715 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a4e39b8d-4a24-486d-897c-39c46b733dec-var-lock\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.250415 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c078273-d5a5-4371-8ad5-091114f61311" (UID: "8c078273-d5a5-4371-8ad5-091114f61311"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.303553 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c078273-d5a5-4371-8ad5-091114f61311-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.684098 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xn5c9" event={"ID":"8c078273-d5a5-4371-8ad5-091114f61311","Type":"ContainerDied","Data":"89e3b97893a5169ad240698afc2fb189a4eb64404cdd443a15a9f6cfb10e9721"} Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.684134 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xn5c9" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.684175 4715 scope.go:117] "RemoveContainer" containerID="4eb9a31a1ec94d641cb6626ecc3ac12dcd709fdb2e8d10b8c871f9e31e25e010" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.684871 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.685134 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.685299 4715 status_manager.go:851] "Failed to get status for pod" podUID="8c078273-d5a5-4371-8ad5-091114f61311" pod="openshift-marketplace/redhat-operators-xn5c9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xn5c9\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.687846 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.688616 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee" exitCode=0 Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.690065 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a4e39b8d-4a24-486d-897c-39c46b733dec","Type":"ContainerDied","Data":"6cace447ac5abeef102610336a403257bfc049925fc91d9ade3e893d261ab743"} Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.690089 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cace447ac5abeef102610336a403257bfc049925fc91d9ade3e893d261ab743" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.690107 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.697612 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.698004 4715 status_manager.go:851] "Failed to get status for pod" podUID="8c078273-d5a5-4371-8ad5-091114f61311" pod="openshift-marketplace/redhat-operators-xn5c9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xn5c9\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.698207 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.707089 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.707399 4715 status_manager.go:851] "Failed to get status for pod" podUID="8c078273-d5a5-4371-8ad5-091114f61311" pod="openshift-marketplace/redhat-operators-xn5c9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xn5c9\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.707747 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.708677 4715 scope.go:117] "RemoveContainer" containerID="f11b5d4926fe1edf270cd26217e42e0d7308218583d6e3bdfeccbcd35be5ff4f" Dec 10 09:38:06 crc kubenswrapper[4715]: I1210 09:38:06.742399 4715 scope.go:117] "RemoveContainer" containerID="89e4fdc15f388f2a2a736ecc0736066e49d6ec159b3056e5d2c94321192cc005" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.607265 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.607902 4715 status_manager.go:851] "Failed to get status for pod" podUID="8c078273-d5a5-4371-8ad5-091114f61311" pod="openshift-marketplace/redhat-operators-xn5c9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xn5c9\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.608685 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.760432 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.761160 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.761823 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.764269 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.764834 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.765084 4715 status_manager.go:851] "Failed to get status for pod" podUID="8c078273-d5a5-4371-8ad5-091114f61311" pod="openshift-marketplace/redhat-operators-xn5c9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xn5c9\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.925094 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.925532 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.925564 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.925775 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.925781 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:38:07 crc kubenswrapper[4715]: I1210 09:38:07.925888 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.027214 4715 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.027243 4715 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.027254 4715 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.702209 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.703042 4715 scope.go:117] "RemoveContainer" containerID="1cdf9dba4b0a7ff1403fb0887a160237b2422a37551c91be9f0a79482750deed" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.703154 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.717776 4715 status_manager.go:851] "Failed to get status for pod" podUID="8c078273-d5a5-4371-8ad5-091114f61311" pod="openshift-marketplace/redhat-operators-xn5c9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xn5c9\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.718061 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.718286 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.718363 4715 scope.go:117] "RemoveContainer" containerID="3d5cd27d023b3814f8052ab46fd7471afe466d858a805a6aac20f2a0580ca335" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.718905 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.733039 4715 scope.go:117] "RemoveContainer" containerID="93f2f29c26e18a7636e1b8f5449244373fc8f8f54c52ab8fccd12e1a9f31b64d" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.745615 4715 scope.go:117] "RemoveContainer" containerID="09d797f9dd9e67afe42057a0986b814175391ad6d817c047373c17124828d338" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.756824 4715 scope.go:117] "RemoveContainer" containerID="d43855a590a0a32bd82ab0272f456428aa2d73512887b3b71681146b4ffdf1ee" Dec 10 09:38:08 crc kubenswrapper[4715]: I1210 09:38:08.772469 4715 scope.go:117] "RemoveContainer" containerID="4e3491cad8cf62ce6f8a664c86c0cbd36c8275c5901673065966958f23e542a9" Dec 10 09:38:09 crc kubenswrapper[4715]: E1210 09:38:09.353522 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="6.4s" Dec 10 09:38:09 crc kubenswrapper[4715]: I1210 09:38:09.612527 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 10 09:38:12 crc kubenswrapper[4715]: E1210 09:38:12.657827 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.201:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187fd118c2403aee openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 09:38:03.220933358 +0000 UTC m=+245.964479619,LastTimestamp:2025-12-10 09:38:03.220933358 +0000 UTC m=+245.964479619,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 09:38:14 crc kubenswrapper[4715]: E1210 09:38:14.136686 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:38:14Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:38:14Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:38:14Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T09:38:14Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:324d8852a972c066be00776701f7ea5f0b533e79431c220f6f6eb0b6b338ace7\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:cb4d03c3e9f89cd7fa16144ef8515fd5e1f3a5505d707667dea76b4e3cc09f9b\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1626095435},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:040ed5bd62bba86f603d3f5f7e1d29d3fab894f04c561f0d97c907340285e79e\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:99471ee5d5867f4757df6da943e7254536ab2a616bf2c62dc6900f0e1ea07109\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1215778122},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:26ea35413bef0c078547a03dc093f1d4f15a7d9fc91f05c687b1a437352f0856\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:ba6f8ed6f9b58e63d979cc9729f5ce2c6f22e99b35ce91b902e198eb78d6b106\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201960779},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1b1026c62413fa239fa4ff6541fe8bda656c1281867ad6ee2c848feccb13c97e\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:2b633ebdc901d19290af4dc2d09e2b59c504c0fc15a3fba410b0ce098e2d5753\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1141987142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:14 crc kubenswrapper[4715]: E1210 09:38:14.137607 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:14 crc kubenswrapper[4715]: E1210 09:38:14.137801 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:14 crc kubenswrapper[4715]: E1210 09:38:14.137986 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:14 crc kubenswrapper[4715]: E1210 09:38:14.138152 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:14 crc kubenswrapper[4715]: E1210 09:38:14.138165 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 09:38:14 crc kubenswrapper[4715]: I1210 09:38:14.604822 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:14 crc kubenswrapper[4715]: I1210 09:38:14.607472 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:14 crc kubenswrapper[4715]: I1210 09:38:14.611390 4715 status_manager.go:851] "Failed to get status for pod" podUID="8c078273-d5a5-4371-8ad5-091114f61311" pod="openshift-marketplace/redhat-operators-xn5c9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xn5c9\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:14 crc kubenswrapper[4715]: I1210 09:38:14.611714 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:14 crc kubenswrapper[4715]: I1210 09:38:14.624004 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ad92614a-879e-4f8e-8066-8b3c6c395ee8" Dec 10 09:38:14 crc kubenswrapper[4715]: I1210 09:38:14.624063 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ad92614a-879e-4f8e-8066-8b3c6c395ee8" Dec 10 09:38:14 crc kubenswrapper[4715]: E1210 09:38:14.624521 4715 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:14 crc kubenswrapper[4715]: I1210 09:38:14.625092 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:14 crc kubenswrapper[4715]: W1210 09:38:14.653543 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-8ee79b90df39902434ad8bce821e1e66add1cf947688afeb854d0b165dc4bfd4 WatchSource:0}: Error finding container 8ee79b90df39902434ad8bce821e1e66add1cf947688afeb854d0b165dc4bfd4: Status 404 returned error can't find the container with id 8ee79b90df39902434ad8bce821e1e66add1cf947688afeb854d0b165dc4bfd4 Dec 10 09:38:14 crc kubenswrapper[4715]: I1210 09:38:14.744543 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8ee79b90df39902434ad8bce821e1e66add1cf947688afeb854d0b165dc4bfd4"} Dec 10 09:38:15 crc kubenswrapper[4715]: I1210 09:38:15.752990 4715 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d287a789aedde268c87eec9cb36697d7176995f0601af2da7d69388f0795ec12" exitCode=0 Dec 10 09:38:15 crc kubenswrapper[4715]: I1210 09:38:15.753047 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d287a789aedde268c87eec9cb36697d7176995f0601af2da7d69388f0795ec12"} Dec 10 09:38:15 crc kubenswrapper[4715]: I1210 09:38:15.753493 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ad92614a-879e-4f8e-8066-8b3c6c395ee8" Dec 10 09:38:15 crc kubenswrapper[4715]: I1210 09:38:15.753531 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ad92614a-879e-4f8e-8066-8b3c6c395ee8" Dec 10 09:38:15 crc kubenswrapper[4715]: E1210 09:38:15.754904 4715 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:15 crc kubenswrapper[4715]: E1210 09:38:15.755017 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.201:6443: connect: connection refused" interval="7s" Dec 10 09:38:15 crc kubenswrapper[4715]: I1210 09:38:15.755014 4715 status_manager.go:851] "Failed to get status for pod" podUID="8c078273-d5a5-4371-8ad5-091114f61311" pod="openshift-marketplace/redhat-operators-xn5c9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xn5c9\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:15 crc kubenswrapper[4715]: I1210 09:38:15.755544 4715 status_manager.go:851] "Failed to get status for pod" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:15 crc kubenswrapper[4715]: I1210 09:38:15.756131 4715 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.201:6443: connect: connection refused" Dec 10 09:38:16 crc kubenswrapper[4715]: I1210 09:38:16.765733 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bf48775e2f5961c23cab8abeeacc07dba2af146893cb88d85a72a15bbc5c11f6"} Dec 10 09:38:16 crc kubenswrapper[4715]: I1210 09:38:16.766237 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f8d5571b366193f117ca831fed59745b970802c2edf73b50127ab30e43add794"} Dec 10 09:38:17 crc kubenswrapper[4715]: I1210 09:38:17.776442 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5367dbb5dde33a25fe69574f1ed62cd75e38d9f1fe6509d139dd29bd945554b2"} Dec 10 09:38:17 crc kubenswrapper[4715]: I1210 09:38:17.776849 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:17 crc kubenswrapper[4715]: I1210 09:38:17.776863 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3d4f015ad907e682f06e54c33f13c4b32c4f49ef2f27217edfc5b38d02cf8c2b"} Dec 10 09:38:17 crc kubenswrapper[4715]: I1210 09:38:17.776876 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"db250b9f610de5e9551f399fc0ee2296501a817762e79f01f486da21d2abd9cc"} Dec 10 09:38:17 crc kubenswrapper[4715]: I1210 09:38:17.776776 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ad92614a-879e-4f8e-8066-8b3c6c395ee8" Dec 10 09:38:17 crc kubenswrapper[4715]: I1210 09:38:17.776894 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ad92614a-879e-4f8e-8066-8b3c6c395ee8" Dec 10 09:38:18 crc kubenswrapper[4715]: I1210 09:38:18.786803 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 10 09:38:18 crc kubenswrapper[4715]: I1210 09:38:18.786854 4715 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667" exitCode=1 Dec 10 09:38:18 crc kubenswrapper[4715]: I1210 09:38:18.786883 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667"} Dec 10 09:38:18 crc kubenswrapper[4715]: I1210 09:38:18.787427 4715 scope.go:117] "RemoveContainer" containerID="b5708a006c4f29c006e96dab8c57b947a12072a8f98afd703cd821a4dc993667" Dec 10 09:38:19 crc kubenswrapper[4715]: I1210 09:38:19.625374 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:19 crc kubenswrapper[4715]: I1210 09:38:19.625896 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:19 crc kubenswrapper[4715]: I1210 09:38:19.633619 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:19 crc kubenswrapper[4715]: I1210 09:38:19.795228 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 10 09:38:19 crc kubenswrapper[4715]: I1210 09:38:19.795388 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5576a432d640cabfb6d164775b81d5f7d279c850bb7ba9aafa40f88fe9f60db7"} Dec 10 09:38:22 crc kubenswrapper[4715]: I1210 09:38:22.819114 4715 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:22 crc kubenswrapper[4715]: I1210 09:38:22.868332 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="929d91bc-107f-44a8-be93-e9507fad8e03" Dec 10 09:38:23 crc kubenswrapper[4715]: I1210 09:38:23.819223 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ad92614a-879e-4f8e-8066-8b3c6c395ee8" Dec 10 09:38:23 crc kubenswrapper[4715]: I1210 09:38:23.819251 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ad92614a-879e-4f8e-8066-8b3c6c395ee8" Dec 10 09:38:23 crc kubenswrapper[4715]: I1210 09:38:23.822895 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:23 crc kubenswrapper[4715]: I1210 09:38:23.824121 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="929d91bc-107f-44a8-be93-e9507fad8e03" Dec 10 09:38:24 crc kubenswrapper[4715]: I1210 09:38:24.825335 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ad92614a-879e-4f8e-8066-8b3c6c395ee8" Dec 10 09:38:24 crc kubenswrapper[4715]: I1210 09:38:24.825379 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ad92614a-879e-4f8e-8066-8b3c6c395ee8" Dec 10 09:38:24 crc kubenswrapper[4715]: I1210 09:38:24.833878 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="929d91bc-107f-44a8-be93-e9507fad8e03" Dec 10 09:38:26 crc kubenswrapper[4715]: I1210 09:38:26.616449 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:38:26 crc kubenswrapper[4715]: I1210 09:38:26.781794 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:38:26 crc kubenswrapper[4715]: I1210 09:38:26.789155 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:38:32 crc kubenswrapper[4715]: I1210 09:38:32.512114 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 09:38:32 crc kubenswrapper[4715]: I1210 09:38:32.588889 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 10 09:38:32 crc kubenswrapper[4715]: I1210 09:38:32.857739 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 09:38:33 crc kubenswrapper[4715]: I1210 09:38:33.527974 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 10 09:38:33 crc kubenswrapper[4715]: I1210 09:38:33.687351 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 10 09:38:33 crc kubenswrapper[4715]: I1210 09:38:33.876532 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.036265 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.072202 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.155446 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.373969 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.533010 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.702398 4715 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.704383 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=32.704362479 podStartE2EDuration="32.704362479s" podCreationTimestamp="2025-12-10 09:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:38:22.905094176 +0000 UTC m=+265.648640427" watchObservedRunningTime="2025-12-10 09:38:34.704362479 +0000 UTC m=+277.447908730" Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.707356 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/redhat-operators-xn5c9"] Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.707406 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.711268 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.722642 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=12.722631988 podStartE2EDuration="12.722631988s" podCreationTimestamp="2025-12-10 09:38:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:38:34.722266537 +0000 UTC m=+277.465812788" watchObservedRunningTime="2025-12-10 09:38:34.722631988 +0000 UTC m=+277.466178239" Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.735244 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 10 09:38:34 crc kubenswrapper[4715]: I1210 09:38:34.818035 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.017386 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.069321 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.357106 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.478642 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.512811 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.613983 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c078273-d5a5-4371-8ad5-091114f61311" path="/var/lib/kubelet/pods/8c078273-d5a5-4371-8ad5-091114f61311/volumes" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.623196 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.642723 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.647206 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.667951 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.939140 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 10 09:38:35 crc kubenswrapper[4715]: I1210 09:38:35.962734 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.020049 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.145785 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.193532 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.204369 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.233079 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.344130 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.366820 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.441754 4715 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.621140 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.637384 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.694478 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.731542 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.733734 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.844448 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.852266 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.932612 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 10 09:38:36 crc kubenswrapper[4715]: I1210 09:38:36.975881 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.007695 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.011196 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.052224 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.074970 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.131383 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.162596 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.173467 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.279099 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.361066 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.386897 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.509827 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.686832 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.744298 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.870604 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 10 09:38:37 crc kubenswrapper[4715]: I1210 09:38:37.876594 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.009424 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.170052 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.267233 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.278204 4715 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.343545 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.349835 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.456254 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.611108 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.707191 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.878196 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.928767 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.970386 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 10 09:38:38 crc kubenswrapper[4715]: I1210 09:38:38.973345 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.095019 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.123143 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.132817 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.267743 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.501206 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.544824 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.592623 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.695595 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.732516 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.802259 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.849381 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.913847 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 10 09:38:39 crc kubenswrapper[4715]: I1210 09:38:39.925005 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.016637 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.045010 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.050869 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.162543 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.219468 4715 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.306991 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.398271 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.456580 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.512684 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.521817 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.601687 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.760626 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.855057 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.896339 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.905838 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 10 09:38:40 crc kubenswrapper[4715]: I1210 09:38:40.987783 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.051674 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.051954 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.074621 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.081556 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.100836 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.112801 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.211500 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.310152 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.337826 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.358393 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.377820 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.575544 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.651749 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.709236 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.740996 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.799326 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 09:38:41 crc kubenswrapper[4715]: I1210 09:38:41.981136 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.028530 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.085275 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.095193 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.183816 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.183891 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.252524 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.318231 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.325368 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.364447 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.389356 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.430624 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.600218 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.644680 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.717177 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.835068 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.876831 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.885642 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 10 09:38:42 crc kubenswrapper[4715]: I1210 09:38:42.906501 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.052640 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.090669 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.115026 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.176711 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.179166 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.297010 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.302705 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.343400 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.347981 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.390198 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.476273 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.481639 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.528905 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.617398 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.840469 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.876534 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 10 09:38:43 crc kubenswrapper[4715]: I1210 09:38:43.987744 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.022662 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.162315 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.220236 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.225532 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.233650 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.251862 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.268471 4715 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.268736 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204" gracePeriod=5 Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.301618 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.332500 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.360827 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.381643 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.401669 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.470814 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.471089 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.544871 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.579074 4715 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.586879 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.598758 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.623693 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.726048 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.806865 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.842291 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 10 09:38:44 crc kubenswrapper[4715]: I1210 09:38:44.920604 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.005409 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.028471 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.107342 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.185348 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.185830 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.203020 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.234182 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.254519 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.339126 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.385212 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.402644 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.438338 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.650795 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.680237 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.730982 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.834745 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.953517 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 10 09:38:45 crc kubenswrapper[4715]: I1210 09:38:45.995758 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.075278 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.085729 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.154282 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.157867 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.162796 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.187737 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.339404 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.340416 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.464220 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.487467 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.521229 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.600268 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.631312 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.684312 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.684978 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.822948 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 10 09:38:46 crc kubenswrapper[4715]: I1210 09:38:46.952080 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.002809 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.042178 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.095060 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.151020 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.319270 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.325849 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.365575 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.380231 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.501404 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.554241 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.584072 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.731828 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.744959 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.795363 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.871004 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.879901 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 10 09:38:47 crc kubenswrapper[4715]: I1210 09:38:47.958371 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 10 09:38:48 crc kubenswrapper[4715]: I1210 09:38:48.008455 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 09:38:48 crc kubenswrapper[4715]: I1210 09:38:48.103285 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 10 09:38:48 crc kubenswrapper[4715]: I1210 09:38:48.175353 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 10 09:38:48 crc kubenswrapper[4715]: I1210 09:38:48.274145 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 10 09:38:48 crc kubenswrapper[4715]: I1210 09:38:48.345679 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 10 09:38:48 crc kubenswrapper[4715]: I1210 09:38:48.351191 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 10 09:38:48 crc kubenswrapper[4715]: I1210 09:38:48.531251 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 09:38:48 crc kubenswrapper[4715]: I1210 09:38:48.706050 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 10 09:38:48 crc kubenswrapper[4715]: I1210 09:38:48.771447 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 10 09:38:48 crc kubenswrapper[4715]: I1210 09:38:48.995263 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.079523 4715 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.116498 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fx8wx"] Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.116776 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fx8wx" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" containerName="registry-server" containerID="cri-o://cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0" gracePeriod=30 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.124325 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vq4cc"] Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.124877 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vq4cc" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerName="registry-server" containerID="cri-o://28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40" gracePeriod=30 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.127027 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.142390 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnj8g"] Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.142635 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" podUID="f37a1d66-5d7d-4d58-9f23-401de6615848" containerName="marketplace-operator" containerID="cri-o://dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67" gracePeriod=30 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.152699 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8t5q2"] Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.153187 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8t5q2" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerName="registry-server" containerID="cri-o://b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64" gracePeriod=30 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.159574 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jrl25"] Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.164400 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jrl25" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" containerName="registry-server" containerID="cri-o://2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea" gracePeriod=30 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.181707 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.182683 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rjxvk"] Dec 10 09:38:49 crc kubenswrapper[4715]: E1210 09:38:49.183076 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c078273-d5a5-4371-8ad5-091114f61311" containerName="extract-utilities" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.183103 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c078273-d5a5-4371-8ad5-091114f61311" containerName="extract-utilities" Dec 10 09:38:49 crc kubenswrapper[4715]: E1210 09:38:49.183117 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.183128 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 09:38:49 crc kubenswrapper[4715]: E1210 09:38:49.183143 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c078273-d5a5-4371-8ad5-091114f61311" containerName="extract-content" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.183151 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c078273-d5a5-4371-8ad5-091114f61311" containerName="extract-content" Dec 10 09:38:49 crc kubenswrapper[4715]: E1210 09:38:49.183160 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c078273-d5a5-4371-8ad5-091114f61311" containerName="registry-server" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.183167 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c078273-d5a5-4371-8ad5-091114f61311" containerName="registry-server" Dec 10 09:38:49 crc kubenswrapper[4715]: E1210 09:38:49.183179 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" containerName="installer" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.183186 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" containerName="installer" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.183295 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.183311 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e39b8d-4a24-486d-897c-39c46b733dec" containerName="installer" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.183330 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c078273-d5a5-4371-8ad5-091114f61311" containerName="registry-server" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.183889 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.196130 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rjxvk"] Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.279470 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.286515 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed4c45ce-d179-4515-8166-9f1c6c7e0913-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rjxvk\" (UID: \"ed4c45ce-d179-4515-8166-9f1c6c7e0913\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.286684 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed4c45ce-d179-4515-8166-9f1c6c7e0913-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rjxvk\" (UID: \"ed4c45ce-d179-4515-8166-9f1c6c7e0913\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.286742 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44pr2\" (UniqueName: \"kubernetes.io/projected/ed4c45ce-d179-4515-8166-9f1c6c7e0913-kube-api-access-44pr2\") pod \"marketplace-operator-79b997595-rjxvk\" (UID: \"ed4c45ce-d179-4515-8166-9f1c6c7e0913\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.388071 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed4c45ce-d179-4515-8166-9f1c6c7e0913-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rjxvk\" (UID: \"ed4c45ce-d179-4515-8166-9f1c6c7e0913\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.388144 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed4c45ce-d179-4515-8166-9f1c6c7e0913-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rjxvk\" (UID: \"ed4c45ce-d179-4515-8166-9f1c6c7e0913\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.388203 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44pr2\" (UniqueName: \"kubernetes.io/projected/ed4c45ce-d179-4515-8166-9f1c6c7e0913-kube-api-access-44pr2\") pod \"marketplace-operator-79b997595-rjxvk\" (UID: \"ed4c45ce-d179-4515-8166-9f1c6c7e0913\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.394726 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ed4c45ce-d179-4515-8166-9f1c6c7e0913-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rjxvk\" (UID: \"ed4c45ce-d179-4515-8166-9f1c6c7e0913\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.395860 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed4c45ce-d179-4515-8166-9f1c6c7e0913-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rjxvk\" (UID: \"ed4c45ce-d179-4515-8166-9f1c6c7e0913\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.410950 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44pr2\" (UniqueName: \"kubernetes.io/projected/ed4c45ce-d179-4515-8166-9f1c6c7e0913-kube-api-access-44pr2\") pod \"marketplace-operator-79b997595-rjxvk\" (UID: \"ed4c45ce-d179-4515-8166-9f1c6c7e0913\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.452150 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.486114 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.579441 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.594290 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.624831 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.667204 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.667855 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.668030 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.672801 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.675585 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.679713 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.698821 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.703710 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.706037 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.734430 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.792811 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p727r\" (UniqueName: \"kubernetes.io/projected/676753c5-d951-4722-a5ef-82e3227e0c7e-kube-api-access-p727r\") pod \"676753c5-d951-4722-a5ef-82e3227e0c7e\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.792851 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-catalog-content\") pod \"676753c5-d951-4722-a5ef-82e3227e0c7e\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.792872 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q7mg\" (UniqueName: \"kubernetes.io/projected/f37a1d66-5d7d-4d58-9f23-401de6615848-kube-api-access-6q7mg\") pod \"f37a1d66-5d7d-4d58-9f23-401de6615848\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.792898 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-utilities\") pod \"676753c5-d951-4722-a5ef-82e3227e0c7e\" (UID: \"676753c5-d951-4722-a5ef-82e3227e0c7e\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.792945 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-catalog-content\") pod \"fa81c629-c83d-444d-a113-de3681b8f0cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.792973 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793008 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-utilities\") pod \"fa81c629-c83d-444d-a113-de3681b8f0cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793031 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-operator-metrics\") pod \"f37a1d66-5d7d-4d58-9f23-401de6615848\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793049 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbjkm\" (UniqueName: \"kubernetes.io/projected/fa81c629-c83d-444d-a113-de3681b8f0cc-kube-api-access-lbjkm\") pod \"fa81c629-c83d-444d-a113-de3681b8f0cc\" (UID: \"fa81c629-c83d-444d-a113-de3681b8f0cc\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793075 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-catalog-content\") pod \"6d6da964-23be-49cc-8555-97edb12b5f28\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793095 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgb4r\" (UniqueName: \"kubernetes.io/projected/078de70a-e498-487c-81a0-e7e3532e7c48-kube-api-access-jgb4r\") pod \"078de70a-e498-487c-81a0-e7e3532e7c48\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793122 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-trusted-ca\") pod \"f37a1d66-5d7d-4d58-9f23-401de6615848\" (UID: \"f37a1d66-5d7d-4d58-9f23-401de6615848\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793150 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793173 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-utilities\") pod \"6d6da964-23be-49cc-8555-97edb12b5f28\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793190 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-utilities\") pod \"078de70a-e498-487c-81a0-e7e3532e7c48\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793210 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-catalog-content\") pod \"078de70a-e498-487c-81a0-e7e3532e7c48\" (UID: \"078de70a-e498-487c-81a0-e7e3532e7c48\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793225 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793243 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793266 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793284 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zp86\" (UniqueName: \"kubernetes.io/projected/6d6da964-23be-49cc-8555-97edb12b5f28-kube-api-access-6zp86\") pod \"6d6da964-23be-49cc-8555-97edb12b5f28\" (UID: \"6d6da964-23be-49cc-8555-97edb12b5f28\") " Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793868 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.793926 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.794006 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.794444 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f37a1d66-5d7d-4d58-9f23-401de6615848" (UID: "f37a1d66-5d7d-4d58-9f23-401de6615848"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.794536 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-utilities" (OuterVolumeSpecName: "utilities") pod "fa81c629-c83d-444d-a113-de3681b8f0cc" (UID: "fa81c629-c83d-444d-a113-de3681b8f0cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.794845 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-utilities" (OuterVolumeSpecName: "utilities") pod "6d6da964-23be-49cc-8555-97edb12b5f28" (UID: "6d6da964-23be-49cc-8555-97edb12b5f28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.795768 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-utilities" (OuterVolumeSpecName: "utilities") pod "078de70a-e498-487c-81a0-e7e3532e7c48" (UID: "078de70a-e498-487c-81a0-e7e3532e7c48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.796390 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.796963 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-utilities" (OuterVolumeSpecName: "utilities") pod "676753c5-d951-4722-a5ef-82e3227e0c7e" (UID: "676753c5-d951-4722-a5ef-82e3227e0c7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.798734 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f37a1d66-5d7d-4d58-9f23-401de6615848" (UID: "f37a1d66-5d7d-4d58-9f23-401de6615848"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.798846 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa81c629-c83d-444d-a113-de3681b8f0cc-kube-api-access-lbjkm" (OuterVolumeSpecName: "kube-api-access-lbjkm") pod "fa81c629-c83d-444d-a113-de3681b8f0cc" (UID: "fa81c629-c83d-444d-a113-de3681b8f0cc"). InnerVolumeSpecName "kube-api-access-lbjkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.800696 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/078de70a-e498-487c-81a0-e7e3532e7c48-kube-api-access-jgb4r" (OuterVolumeSpecName: "kube-api-access-jgb4r") pod "078de70a-e498-487c-81a0-e7e3532e7c48" (UID: "078de70a-e498-487c-81a0-e7e3532e7c48"). InnerVolumeSpecName "kube-api-access-jgb4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.800962 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f37a1d66-5d7d-4d58-9f23-401de6615848-kube-api-access-6q7mg" (OuterVolumeSpecName: "kube-api-access-6q7mg") pod "f37a1d66-5d7d-4d58-9f23-401de6615848" (UID: "f37a1d66-5d7d-4d58-9f23-401de6615848"). InnerVolumeSpecName "kube-api-access-6q7mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.801034 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.802799 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d6da964-23be-49cc-8555-97edb12b5f28-kube-api-access-6zp86" (OuterVolumeSpecName: "kube-api-access-6zp86") pod "6d6da964-23be-49cc-8555-97edb12b5f28" (UID: "6d6da964-23be-49cc-8555-97edb12b5f28"). InnerVolumeSpecName "kube-api-access-6zp86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.807228 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/676753c5-d951-4722-a5ef-82e3227e0c7e-kube-api-access-p727r" (OuterVolumeSpecName: "kube-api-access-p727r") pod "676753c5-d951-4722-a5ef-82e3227e0c7e" (UID: "676753c5-d951-4722-a5ef-82e3227e0c7e"). InnerVolumeSpecName "kube-api-access-p727r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.835264 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "676753c5-d951-4722-a5ef-82e3227e0c7e" (UID: "676753c5-d951-4722-a5ef-82e3227e0c7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.864826 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "078de70a-e498-487c-81a0-e7e3532e7c48" (UID: "078de70a-e498-487c-81a0-e7e3532e7c48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.867887 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa81c629-c83d-444d-a113-de3681b8f0cc" (UID: "fa81c629-c83d-444d-a113-de3681b8f0cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897011 4715 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897270 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897363 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897432 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbjkm\" (UniqueName: \"kubernetes.io/projected/fa81c629-c83d-444d-a113-de3681b8f0cc-kube-api-access-lbjkm\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897489 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgb4r\" (UniqueName: \"kubernetes.io/projected/078de70a-e498-487c-81a0-e7e3532e7c48-kube-api-access-jgb4r\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897545 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f37a1d66-5d7d-4d58-9f23-401de6615848-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897610 4715 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897673 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897733 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897793 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078de70a-e498-487c-81a0-e7e3532e7c48-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897847 4715 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897900 4715 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.897992 4715 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.898050 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zp86\" (UniqueName: \"kubernetes.io/projected/6d6da964-23be-49cc-8555-97edb12b5f28-kube-api-access-6zp86\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.898105 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p727r\" (UniqueName: \"kubernetes.io/projected/676753c5-d951-4722-a5ef-82e3227e0c7e-kube-api-access-p727r\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.898169 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.898225 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q7mg\" (UniqueName: \"kubernetes.io/projected/f37a1d66-5d7d-4d58-9f23-401de6615848-kube-api-access-6q7mg\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.898280 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/676753c5-d951-4722-a5ef-82e3227e0c7e-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.898365 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa81c629-c83d-444d-a113-de3681b8f0cc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.941148 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d6da964-23be-49cc-8555-97edb12b5f28" (UID: "6d6da964-23be-49cc-8555-97edb12b5f28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.953290 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.963175 4715 generic.go:334] "Generic (PLEG): container finished" podID="078de70a-e498-487c-81a0-e7e3532e7c48" containerID="cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0" exitCode=0 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.963247 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fx8wx" event={"ID":"078de70a-e498-487c-81a0-e7e3532e7c48","Type":"ContainerDied","Data":"cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0"} Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.963259 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fx8wx" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.963276 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fx8wx" event={"ID":"078de70a-e498-487c-81a0-e7e3532e7c48","Type":"ContainerDied","Data":"1c89e1c0681d3392c630569a2783cc732274983b4b44a0088b0693a46706e275"} Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.963298 4715 scope.go:117] "RemoveContainer" containerID="cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.965158 4715 generic.go:334] "Generic (PLEG): container finished" podID="f37a1d66-5d7d-4d58-9f23-401de6615848" containerID="dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67" exitCode=0 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.965197 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.965230 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" event={"ID":"f37a1d66-5d7d-4d58-9f23-401de6615848","Type":"ContainerDied","Data":"dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67"} Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.965260 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dnj8g" event={"ID":"f37a1d66-5d7d-4d58-9f23-401de6615848","Type":"ContainerDied","Data":"48b484192210a42345e77cd3119ac79f0772d799c7d0f250694d47912cdf42b1"} Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.967539 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.967590 4715 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204" exitCode=137 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.967682 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.971836 4715 generic.go:334] "Generic (PLEG): container finished" podID="6d6da964-23be-49cc-8555-97edb12b5f28" containerID="2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea" exitCode=0 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.971897 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrl25" event={"ID":"6d6da964-23be-49cc-8555-97edb12b5f28","Type":"ContainerDied","Data":"2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea"} Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.971938 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jrl25" event={"ID":"6d6da964-23be-49cc-8555-97edb12b5f28","Type":"ContainerDied","Data":"344c2e1defc32a94cd0552d88ecdd5a6cfd6cbf3486ea43452d5f19a85500103"} Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.971999 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jrl25" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.978870 4715 scope.go:117] "RemoveContainer" containerID="80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.978862 4715 generic.go:334] "Generic (PLEG): container finished" podID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerID="b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64" exitCode=0 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.978908 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t5q2" event={"ID":"676753c5-d951-4722-a5ef-82e3227e0c7e","Type":"ContainerDied","Data":"b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64"} Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.979107 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8t5q2" event={"ID":"676753c5-d951-4722-a5ef-82e3227e0c7e","Type":"ContainerDied","Data":"77f7460c587d9dee9029c21e47756c323a831ed5bf2bb0fbec71bff2c409cbfe"} Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.980825 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8t5q2" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.981821 4715 generic.go:334] "Generic (PLEG): container finished" podID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerID="28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40" exitCode=0 Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.981861 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq4cc" event={"ID":"fa81c629-c83d-444d-a113-de3681b8f0cc","Type":"ContainerDied","Data":"28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40"} Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.981887 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vq4cc" Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.981891 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vq4cc" event={"ID":"fa81c629-c83d-444d-a113-de3681b8f0cc","Type":"ContainerDied","Data":"92be6d23c3789901906037041ddc11a0130c3415dd69608d9b8e3311fe81be39"} Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.995280 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnj8g"] Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.998604 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dnj8g"] Dec 10 09:38:49 crc kubenswrapper[4715]: I1210 09:38:49.999240 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d6da964-23be-49cc-8555-97edb12b5f28-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.000611 4715 scope.go:117] "RemoveContainer" containerID="e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.011978 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fx8wx"] Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.017036 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fx8wx"] Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.026753 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8t5q2"] Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.030809 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8t5q2"] Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.033831 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jrl25"] Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.037161 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jrl25"] Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.041212 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vq4cc"] Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.044280 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vq4cc"] Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.047374 4715 scope.go:117] "RemoveContainer" containerID="cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.047937 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0\": container with ID starting with cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0 not found: ID does not exist" containerID="cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.047980 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0"} err="failed to get container status \"cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0\": rpc error: code = NotFound desc = could not find container \"cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0\": container with ID starting with cbfb87109666978d701ff5434b5b45f204f98dd4d5a1b52f35a7f84d10505ea0 not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.048012 4715 scope.go:117] "RemoveContainer" containerID="80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.048873 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e\": container with ID starting with 80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e not found: ID does not exist" containerID="80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.048901 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e"} err="failed to get container status \"80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e\": rpc error: code = NotFound desc = could not find container \"80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e\": container with ID starting with 80a9fa1383dccf4cfe5284a21de88a0b813ae9d0dba4acaa4efe4ed90f29d19e not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.048933 4715 scope.go:117] "RemoveContainer" containerID="e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.049454 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d\": container with ID starting with e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d not found: ID does not exist" containerID="e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.049482 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d"} err="failed to get container status \"e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d\": rpc error: code = NotFound desc = could not find container \"e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d\": container with ID starting with e78df46ec9a660cd3e9148d9a512e9f823d084482152eb1b06af5bf177e61d5d not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.049501 4715 scope.go:117] "RemoveContainer" containerID="dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.061931 4715 scope.go:117] "RemoveContainer" containerID="dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.062292 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67\": container with ID starting with dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67 not found: ID does not exist" containerID="dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.062330 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67"} err="failed to get container status \"dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67\": rpc error: code = NotFound desc = could not find container \"dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67\": container with ID starting with dd5f11bf2f54b07754e276e3d3f364ae7d02e09c530cfb0bf7ca2620035d1c67 not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.062358 4715 scope.go:117] "RemoveContainer" containerID="6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.072330 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.073103 4715 scope.go:117] "RemoveContainer" containerID="6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.073397 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204\": container with ID starting with 6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204 not found: ID does not exist" containerID="6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.073432 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204"} err="failed to get container status \"6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204\": rpc error: code = NotFound desc = could not find container \"6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204\": container with ID starting with 6a756a4c1ea60a2a48ffb6095517563c2112cd838b43df68f9e7b01789d26204 not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.073457 4715 scope.go:117] "RemoveContainer" containerID="2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.088343 4715 scope.go:117] "RemoveContainer" containerID="c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.106711 4715 scope.go:117] "RemoveContainer" containerID="97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.123059 4715 scope.go:117] "RemoveContainer" containerID="2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.123524 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea\": container with ID starting with 2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea not found: ID does not exist" containerID="2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.123549 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea"} err="failed to get container status \"2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea\": rpc error: code = NotFound desc = could not find container \"2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea\": container with ID starting with 2b30d14185450476ba88790a1a0f06c4bae72d0110e238e8d240b0cbab774aea not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.123577 4715 scope.go:117] "RemoveContainer" containerID="c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.124048 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7\": container with ID starting with c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7 not found: ID does not exist" containerID="c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.124095 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7"} err="failed to get container status \"c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7\": rpc error: code = NotFound desc = could not find container \"c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7\": container with ID starting with c81239cb2228b88a4560b30db1cbd517316af6b9f3bc2fd9a73323cb321310d7 not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.124181 4715 scope.go:117] "RemoveContainer" containerID="97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.124481 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e\": container with ID starting with 97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e not found: ID does not exist" containerID="97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.124501 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e"} err="failed to get container status \"97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e\": rpc error: code = NotFound desc = could not find container \"97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e\": container with ID starting with 97a241f2484fa4bd64bacaa244b848d9f3a106f39fe665a5220293d61fe0288e not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.124514 4715 scope.go:117] "RemoveContainer" containerID="b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.142400 4715 scope.go:117] "RemoveContainer" containerID="e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.162322 4715 scope.go:117] "RemoveContainer" containerID="cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.187759 4715 scope.go:117] "RemoveContainer" containerID="b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.188428 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64\": container with ID starting with b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64 not found: ID does not exist" containerID="b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.188462 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64"} err="failed to get container status \"b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64\": rpc error: code = NotFound desc = could not find container \"b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64\": container with ID starting with b4386d553e415acc0ff1d4bd8b2a0d3341fa3a925796619e6876ee294b552c64 not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.188492 4715 scope.go:117] "RemoveContainer" containerID="e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.188928 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc\": container with ID starting with e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc not found: ID does not exist" containerID="e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.188954 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc"} err="failed to get container status \"e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc\": rpc error: code = NotFound desc = could not find container \"e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc\": container with ID starting with e570034e6ca3486dfff3415ab4da6b41b1fa73a192c42e460da37b75baed3edc not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.188972 4715 scope.go:117] "RemoveContainer" containerID="cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.189308 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c\": container with ID starting with cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c not found: ID does not exist" containerID="cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.189339 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c"} err="failed to get container status \"cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c\": rpc error: code = NotFound desc = could not find container \"cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c\": container with ID starting with cb088aab9b4ad4e8b80b789e3d02df17663ce7b0f4bd5bf4dcd936495316574c not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.189355 4715 scope.go:117] "RemoveContainer" containerID="28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.200634 4715 scope.go:117] "RemoveContainer" containerID="469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.213184 4715 scope.go:117] "RemoveContainer" containerID="baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.231964 4715 scope.go:117] "RemoveContainer" containerID="28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.232347 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40\": container with ID starting with 28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40 not found: ID does not exist" containerID="28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.232393 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40"} err="failed to get container status \"28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40\": rpc error: code = NotFound desc = could not find container \"28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40\": container with ID starting with 28b76bf096ddd7c89230d59073c7b2a377c24b12829cfa809ba41cb6ff611a40 not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.232427 4715 scope.go:117] "RemoveContainer" containerID="469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.232854 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87\": container with ID starting with 469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87 not found: ID does not exist" containerID="469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.232895 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87"} err="failed to get container status \"469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87\": rpc error: code = NotFound desc = could not find container \"469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87\": container with ID starting with 469e2fb833b1164ba17a3cdde9f5a8040a365038765277f87bb03f20f4373b87 not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.232942 4715 scope.go:117] "RemoveContainer" containerID="baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da" Dec 10 09:38:50 crc kubenswrapper[4715]: E1210 09:38:50.233345 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da\": container with ID starting with baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da not found: ID does not exist" containerID="baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.233398 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da"} err="failed to get container status \"baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da\": rpc error: code = NotFound desc = could not find container \"baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da\": container with ID starting with baa5dcb7bce102dd6d6a335d8947def0a443f5f0cc0306e51a1cf736a48100da not found: ID does not exist" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.368542 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.565358 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.717490 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rjxvk"] Dec 10 09:38:50 crc kubenswrapper[4715]: I1210 09:38:50.989026 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" event={"ID":"ed4c45ce-d179-4515-8166-9f1c6c7e0913","Type":"ContainerStarted","Data":"453821e086dd744a583cce696b210905b1a7fac14b7cab4c558b0807520086bc"} Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.002548 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.552424 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.612064 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" path="/var/lib/kubelet/pods/078de70a-e498-487c-81a0-e7e3532e7c48/volumes" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.613252 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" path="/var/lib/kubelet/pods/676753c5-d951-4722-a5ef-82e3227e0c7e/volumes" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.613855 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" path="/var/lib/kubelet/pods/6d6da964-23be-49cc-8555-97edb12b5f28/volumes" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.614888 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f37a1d66-5d7d-4d58-9f23-401de6615848" path="/var/lib/kubelet/pods/f37a1d66-5d7d-4d58-9f23-401de6615848/volumes" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.615265 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.615764 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" path="/var/lib/kubelet/pods/fa81c629-c83d-444d-a113-de3681b8f0cc/volumes" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.616582 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.625687 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.625731 4715 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="76ccffbd-6c86-4117-870d-383b38643a87" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.628965 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.628995 4715 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="76ccffbd-6c86-4117-870d-383b38643a87" Dec 10 09:38:51 crc kubenswrapper[4715]: I1210 09:38:51.884824 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 09:38:52 crc kubenswrapper[4715]: I1210 09:38:52.003625 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" event={"ID":"ed4c45ce-d179-4515-8166-9f1c6c7e0913","Type":"ContainerStarted","Data":"ac27120c734ef24260975a07a009e0d6f16996c25446d5997e50632db7253fde"} Dec 10 09:38:52 crc kubenswrapper[4715]: I1210 09:38:52.003868 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:52 crc kubenswrapper[4715]: I1210 09:38:52.007158 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" Dec 10 09:38:52 crc kubenswrapper[4715]: I1210 09:38:52.023083 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rjxvk" podStartSLOduration=3.023062029 podStartE2EDuration="3.023062029s" podCreationTimestamp="2025-12-10 09:38:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:38:52.02044859 +0000 UTC m=+294.763994841" watchObservedRunningTime="2025-12-10 09:38:52.023062029 +0000 UTC m=+294.766608270" Dec 10 09:38:57 crc kubenswrapper[4715]: I1210 09:38:57.441730 4715 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 10 09:39:19 crc kubenswrapper[4715]: I1210 09:39:19.098642 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7qnsk"] Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106123 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sc95c"] Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106357 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerName="extract-utilities" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106371 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerName="extract-utilities" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106379 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" containerName="extract-utilities" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106385 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" containerName="extract-utilities" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106395 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerName="extract-utilities" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106401 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerName="extract-utilities" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106410 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" containerName="extract-utilities" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106417 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" containerName="extract-utilities" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106424 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106429 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106437 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f37a1d66-5d7d-4d58-9f23-401de6615848" containerName="marketplace-operator" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106442 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f37a1d66-5d7d-4d58-9f23-401de6615848" containerName="marketplace-operator" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106450 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerName="extract-content" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106457 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerName="extract-content" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106464 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerName="extract-content" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106472 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerName="extract-content" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106481 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106487 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106496 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106503 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106512 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106518 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106526 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" containerName="extract-content" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106532 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" containerName="extract-content" Dec 10 09:39:20 crc kubenswrapper[4715]: E1210 09:39:20.106541 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" containerName="extract-content" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106547 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" containerName="extract-content" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106646 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f37a1d66-5d7d-4d58-9f23-401de6615848" containerName="marketplace-operator" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106659 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="078de70a-e498-487c-81a0-e7e3532e7c48" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106670 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d6da964-23be-49cc-8555-97edb12b5f28" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106677 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="676753c5-d951-4722-a5ef-82e3227e0c7e" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.106688 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa81c629-c83d-444d-a113-de3681b8f0cc" containerName="registry-server" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.107387 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.109295 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.120301 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sc95c"] Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.212168 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-catalog-content\") pod \"redhat-marketplace-sc95c\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.212275 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-utilities\") pod \"redhat-marketplace-sc95c\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.212328 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbl6w\" (UniqueName: \"kubernetes.io/projected/434fa3ff-6093-4154-82e9-e0612d48cea0-kube-api-access-mbl6w\") pod \"redhat-marketplace-sc95c\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.310164 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nv8h7"] Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.311173 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.313466 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-utilities\") pod \"redhat-marketplace-sc95c\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.313521 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbl6w\" (UniqueName: \"kubernetes.io/projected/434fa3ff-6093-4154-82e9-e0612d48cea0-kube-api-access-mbl6w\") pod \"redhat-marketplace-sc95c\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.313611 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-catalog-content\") pod \"redhat-marketplace-sc95c\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.313990 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-utilities\") pod \"redhat-marketplace-sc95c\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.314182 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-catalog-content\") pod \"redhat-marketplace-sc95c\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.315888 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.321653 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nv8h7"] Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.344399 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbl6w\" (UniqueName: \"kubernetes.io/projected/434fa3ff-6093-4154-82e9-e0612d48cea0-kube-api-access-mbl6w\") pod \"redhat-marketplace-sc95c\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.415313 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h67c5\" (UniqueName: \"kubernetes.io/projected/956621e9-dfe5-4b87-97c8-458b6f3c5b6d-kube-api-access-h67c5\") pod \"community-operators-nv8h7\" (UID: \"956621e9-dfe5-4b87-97c8-458b6f3c5b6d\") " pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.415473 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/956621e9-dfe5-4b87-97c8-458b6f3c5b6d-catalog-content\") pod \"community-operators-nv8h7\" (UID: \"956621e9-dfe5-4b87-97c8-458b6f3c5b6d\") " pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.415609 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/956621e9-dfe5-4b87-97c8-458b6f3c5b6d-utilities\") pod \"community-operators-nv8h7\" (UID: \"956621e9-dfe5-4b87-97c8-458b6f3c5b6d\") " pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.423619 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.517603 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/956621e9-dfe5-4b87-97c8-458b6f3c5b6d-utilities\") pod \"community-operators-nv8h7\" (UID: \"956621e9-dfe5-4b87-97c8-458b6f3c5b6d\") " pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.517892 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h67c5\" (UniqueName: \"kubernetes.io/projected/956621e9-dfe5-4b87-97c8-458b6f3c5b6d-kube-api-access-h67c5\") pod \"community-operators-nv8h7\" (UID: \"956621e9-dfe5-4b87-97c8-458b6f3c5b6d\") " pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.517956 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/956621e9-dfe5-4b87-97c8-458b6f3c5b6d-catalog-content\") pod \"community-operators-nv8h7\" (UID: \"956621e9-dfe5-4b87-97c8-458b6f3c5b6d\") " pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.518308 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/956621e9-dfe5-4b87-97c8-458b6f3c5b6d-utilities\") pod \"community-operators-nv8h7\" (UID: \"956621e9-dfe5-4b87-97c8-458b6f3c5b6d\") " pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.518481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/956621e9-dfe5-4b87-97c8-458b6f3c5b6d-catalog-content\") pod \"community-operators-nv8h7\" (UID: \"956621e9-dfe5-4b87-97c8-458b6f3c5b6d\") " pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.536781 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h67c5\" (UniqueName: \"kubernetes.io/projected/956621e9-dfe5-4b87-97c8-458b6f3c5b6d-kube-api-access-h67c5\") pod \"community-operators-nv8h7\" (UID: \"956621e9-dfe5-4b87-97c8-458b6f3c5b6d\") " pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.637978 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.812372 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nv8h7"] Dec 10 09:39:20 crc kubenswrapper[4715]: I1210 09:39:20.816289 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sc95c"] Dec 10 09:39:20 crc kubenswrapper[4715]: W1210 09:39:20.818368 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod434fa3ff_6093_4154_82e9_e0612d48cea0.slice/crio-178ec0d3317b8eff0a855e3ffcf61482f8fea67f39000126d6a809791d063f91 WatchSource:0}: Error finding container 178ec0d3317b8eff0a855e3ffcf61482f8fea67f39000126d6a809791d063f91: Status 404 returned error can't find the container with id 178ec0d3317b8eff0a855e3ffcf61482f8fea67f39000126d6a809791d063f91 Dec 10 09:39:21 crc kubenswrapper[4715]: I1210 09:39:21.159676 4715 generic.go:334] "Generic (PLEG): container finished" podID="956621e9-dfe5-4b87-97c8-458b6f3c5b6d" containerID="1a151f1c8be5cca955dfb95cfe63a70dca5e098b0d25399af18e9fe204996df0" exitCode=0 Dec 10 09:39:21 crc kubenswrapper[4715]: I1210 09:39:21.159853 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nv8h7" event={"ID":"956621e9-dfe5-4b87-97c8-458b6f3c5b6d","Type":"ContainerDied","Data":"1a151f1c8be5cca955dfb95cfe63a70dca5e098b0d25399af18e9fe204996df0"} Dec 10 09:39:21 crc kubenswrapper[4715]: I1210 09:39:21.160149 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nv8h7" event={"ID":"956621e9-dfe5-4b87-97c8-458b6f3c5b6d","Type":"ContainerStarted","Data":"e9eded282d60fe049f1f8686ac4323a9044ef573bbb93aeac2666bddf0ec089f"} Dec 10 09:39:21 crc kubenswrapper[4715]: I1210 09:39:21.163388 4715 generic.go:334] "Generic (PLEG): container finished" podID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerID="6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038" exitCode=0 Dec 10 09:39:21 crc kubenswrapper[4715]: I1210 09:39:21.163433 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sc95c" event={"ID":"434fa3ff-6093-4154-82e9-e0612d48cea0","Type":"ContainerDied","Data":"6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038"} Dec 10 09:39:21 crc kubenswrapper[4715]: I1210 09:39:21.163463 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sc95c" event={"ID":"434fa3ff-6093-4154-82e9-e0612d48cea0","Type":"ContainerStarted","Data":"178ec0d3317b8eff0a855e3ffcf61482f8fea67f39000126d6a809791d063f91"} Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.704341 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hb8cn"] Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.706249 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.711256 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.715929 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hb8cn"] Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.847439 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a49347-3b45-4bd9-a158-6deb79a28d2d-utilities\") pod \"redhat-operators-hb8cn\" (UID: \"63a49347-3b45-4bd9-a158-6deb79a28d2d\") " pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.847515 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j8zr\" (UniqueName: \"kubernetes.io/projected/63a49347-3b45-4bd9-a158-6deb79a28d2d-kube-api-access-7j8zr\") pod \"redhat-operators-hb8cn\" (UID: \"63a49347-3b45-4bd9-a158-6deb79a28d2d\") " pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.847547 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a49347-3b45-4bd9-a158-6deb79a28d2d-catalog-content\") pod \"redhat-operators-hb8cn\" (UID: \"63a49347-3b45-4bd9-a158-6deb79a28d2d\") " pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.901074 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tjntt"] Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.902003 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.904371 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.914157 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tjntt"] Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.949489 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn5wp\" (UniqueName: \"kubernetes.io/projected/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-kube-api-access-bn5wp\") pod \"certified-operators-tjntt\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.949543 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-utilities\") pod \"certified-operators-tjntt\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.949592 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a49347-3b45-4bd9-a158-6deb79a28d2d-utilities\") pod \"redhat-operators-hb8cn\" (UID: \"63a49347-3b45-4bd9-a158-6deb79a28d2d\") " pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.949662 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j8zr\" (UniqueName: \"kubernetes.io/projected/63a49347-3b45-4bd9-a158-6deb79a28d2d-kube-api-access-7j8zr\") pod \"redhat-operators-hb8cn\" (UID: \"63a49347-3b45-4bd9-a158-6deb79a28d2d\") " pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.949690 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-catalog-content\") pod \"certified-operators-tjntt\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.949709 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a49347-3b45-4bd9-a158-6deb79a28d2d-catalog-content\") pod \"redhat-operators-hb8cn\" (UID: \"63a49347-3b45-4bd9-a158-6deb79a28d2d\") " pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.950207 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a49347-3b45-4bd9-a158-6deb79a28d2d-catalog-content\") pod \"redhat-operators-hb8cn\" (UID: \"63a49347-3b45-4bd9-a158-6deb79a28d2d\") " pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.951622 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a49347-3b45-4bd9-a158-6deb79a28d2d-utilities\") pod \"redhat-operators-hb8cn\" (UID: \"63a49347-3b45-4bd9-a158-6deb79a28d2d\") " pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:22 crc kubenswrapper[4715]: I1210 09:39:22.968031 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j8zr\" (UniqueName: \"kubernetes.io/projected/63a49347-3b45-4bd9-a158-6deb79a28d2d-kube-api-access-7j8zr\") pod \"redhat-operators-hb8cn\" (UID: \"63a49347-3b45-4bd9-a158-6deb79a28d2d\") " pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.023425 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.050442 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-catalog-content\") pod \"certified-operators-tjntt\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.050912 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn5wp\" (UniqueName: \"kubernetes.io/projected/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-kube-api-access-bn5wp\") pod \"certified-operators-tjntt\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.051029 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-utilities\") pod \"certified-operators-tjntt\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.051248 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-catalog-content\") pod \"certified-operators-tjntt\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.051538 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-utilities\") pod \"certified-operators-tjntt\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.070114 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn5wp\" (UniqueName: \"kubernetes.io/projected/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-kube-api-access-bn5wp\") pod \"certified-operators-tjntt\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.174327 4715 generic.go:334] "Generic (PLEG): container finished" podID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerID="9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179" exitCode=0 Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.174380 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sc95c" event={"ID":"434fa3ff-6093-4154-82e9-e0612d48cea0","Type":"ContainerDied","Data":"9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179"} Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.177406 4715 generic.go:334] "Generic (PLEG): container finished" podID="956621e9-dfe5-4b87-97c8-458b6f3c5b6d" containerID="2e680d676cdb8136f2fd87a956175e662f50775c09563204b4cb4daf2f8526c6" exitCode=0 Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.177447 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nv8h7" event={"ID":"956621e9-dfe5-4b87-97c8-458b6f3c5b6d","Type":"ContainerDied","Data":"2e680d676cdb8136f2fd87a956175e662f50775c09563204b4cb4daf2f8526c6"} Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.213974 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.253771 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hb8cn"] Dec 10 09:39:23 crc kubenswrapper[4715]: W1210 09:39:23.261048 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63a49347_3b45_4bd9_a158_6deb79a28d2d.slice/crio-45105cb5000294af04cf975bfde0994373c093381d88d151d686d583ba85b6d3 WatchSource:0}: Error finding container 45105cb5000294af04cf975bfde0994373c093381d88d151d686d583ba85b6d3: Status 404 returned error can't find the container with id 45105cb5000294af04cf975bfde0994373c093381d88d151d686d583ba85b6d3 Dec 10 09:39:23 crc kubenswrapper[4715]: I1210 09:39:23.369785 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tjntt"] Dec 10 09:39:23 crc kubenswrapper[4715]: W1210 09:39:23.372218 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e744f3a_62da_4f4e_a7ae_5fec5c47873f.slice/crio-9e7ef8efe9e13ea8d9e027e850dd280c759b228eed467de182afb4649906bccb WatchSource:0}: Error finding container 9e7ef8efe9e13ea8d9e027e850dd280c759b228eed467de182afb4649906bccb: Status 404 returned error can't find the container with id 9e7ef8efe9e13ea8d9e027e850dd280c759b228eed467de182afb4649906bccb Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.024575 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-j2l56"] Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.024876 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" podUID="eff08ce1-5a34-48f6-9f72-1e548f563b20" containerName="controller-manager" containerID="cri-o://b5aad762bd37a1e7dc07b3550cd4a1bc33d28a93d1f23cffb1a73b6f40403908" gracePeriod=30 Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.128498 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl"] Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.128729 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" podUID="526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" containerName="route-controller-manager" containerID="cri-o://2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62" gracePeriod=30 Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.187557 4715 generic.go:334] "Generic (PLEG): container finished" podID="63a49347-3b45-4bd9-a158-6deb79a28d2d" containerID="c3e454083cb14ca680eb3e193a7daa822dba1bee1aec621a9fd44f44db7d5fb3" exitCode=0 Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.187679 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb8cn" event={"ID":"63a49347-3b45-4bd9-a158-6deb79a28d2d","Type":"ContainerDied","Data":"c3e454083cb14ca680eb3e193a7daa822dba1bee1aec621a9fd44f44db7d5fb3"} Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.187726 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb8cn" event={"ID":"63a49347-3b45-4bd9-a158-6deb79a28d2d","Type":"ContainerStarted","Data":"45105cb5000294af04cf975bfde0994373c093381d88d151d686d583ba85b6d3"} Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.195920 4715 generic.go:334] "Generic (PLEG): container finished" podID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerID="7a5e1aa3f73ff0a3a8cdba61fce3c35b91d6dd0715667eeacab6bacfdb3a1fcd" exitCode=0 Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.196123 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjntt" event={"ID":"4e744f3a-62da-4f4e-a7ae-5fec5c47873f","Type":"ContainerDied","Data":"7a5e1aa3f73ff0a3a8cdba61fce3c35b91d6dd0715667eeacab6bacfdb3a1fcd"} Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.196267 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjntt" event={"ID":"4e744f3a-62da-4f4e-a7ae-5fec5c47873f","Type":"ContainerStarted","Data":"9e7ef8efe9e13ea8d9e027e850dd280c759b228eed467de182afb4649906bccb"} Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.198360 4715 generic.go:334] "Generic (PLEG): container finished" podID="eff08ce1-5a34-48f6-9f72-1e548f563b20" containerID="b5aad762bd37a1e7dc07b3550cd4a1bc33d28a93d1f23cffb1a73b6f40403908" exitCode=0 Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.198438 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" event={"ID":"eff08ce1-5a34-48f6-9f72-1e548f563b20","Type":"ContainerDied","Data":"b5aad762bd37a1e7dc07b3550cd4a1bc33d28a93d1f23cffb1a73b6f40403908"} Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.203628 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nv8h7" event={"ID":"956621e9-dfe5-4b87-97c8-458b6f3c5b6d","Type":"ContainerStarted","Data":"bfbabc271f6e23554028d3aa6388ba2e2a92c012dcf10fff1f17b15d71d7c429"} Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.213430 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sc95c" event={"ID":"434fa3ff-6093-4154-82e9-e0612d48cea0","Type":"ContainerStarted","Data":"7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a"} Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.241366 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nv8h7" podStartSLOduration=1.5013418299999999 podStartE2EDuration="4.241350319s" podCreationTimestamp="2025-12-10 09:39:20 +0000 UTC" firstStartedPulling="2025-12-10 09:39:21.162231241 +0000 UTC m=+323.905777492" lastFinishedPulling="2025-12-10 09:39:23.90223973 +0000 UTC m=+326.645785981" observedRunningTime="2025-12-10 09:39:24.238966189 +0000 UTC m=+326.982512440" watchObservedRunningTime="2025-12-10 09:39:24.241350319 +0000 UTC m=+326.984896570" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.276578 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sc95c" podStartSLOduration=1.642671787 podStartE2EDuration="4.276553431s" podCreationTimestamp="2025-12-10 09:39:20 +0000 UTC" firstStartedPulling="2025-12-10 09:39:21.164971604 +0000 UTC m=+323.908517855" lastFinishedPulling="2025-12-10 09:39:23.798853248 +0000 UTC m=+326.542399499" observedRunningTime="2025-12-10 09:39:24.270955061 +0000 UTC m=+327.014501332" watchObservedRunningTime="2025-12-10 09:39:24.276553431 +0000 UTC m=+327.020099682" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.375039 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.575292 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwz7w\" (UniqueName: \"kubernetes.io/projected/eff08ce1-5a34-48f6-9f72-1e548f563b20-kube-api-access-gwz7w\") pod \"eff08ce1-5a34-48f6-9f72-1e548f563b20\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.575345 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eff08ce1-5a34-48f6-9f72-1e548f563b20-serving-cert\") pod \"eff08ce1-5a34-48f6-9f72-1e548f563b20\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.575378 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-config\") pod \"eff08ce1-5a34-48f6-9f72-1e548f563b20\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.575396 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-proxy-ca-bundles\") pod \"eff08ce1-5a34-48f6-9f72-1e548f563b20\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.575414 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-client-ca\") pod \"eff08ce1-5a34-48f6-9f72-1e548f563b20\" (UID: \"eff08ce1-5a34-48f6-9f72-1e548f563b20\") " Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.576355 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "eff08ce1-5a34-48f6-9f72-1e548f563b20" (UID: "eff08ce1-5a34-48f6-9f72-1e548f563b20"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.576383 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-config" (OuterVolumeSpecName: "config") pod "eff08ce1-5a34-48f6-9f72-1e548f563b20" (UID: "eff08ce1-5a34-48f6-9f72-1e548f563b20"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.576395 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-client-ca" (OuterVolumeSpecName: "client-ca") pod "eff08ce1-5a34-48f6-9f72-1e548f563b20" (UID: "eff08ce1-5a34-48f6-9f72-1e548f563b20"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.581112 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff08ce1-5a34-48f6-9f72-1e548f563b20-kube-api-access-gwz7w" (OuterVolumeSpecName: "kube-api-access-gwz7w") pod "eff08ce1-5a34-48f6-9f72-1e548f563b20" (UID: "eff08ce1-5a34-48f6-9f72-1e548f563b20"). InnerVolumeSpecName "kube-api-access-gwz7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.583686 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.585130 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eff08ce1-5a34-48f6-9f72-1e548f563b20-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "eff08ce1-5a34-48f6-9f72-1e548f563b20" (UID: "eff08ce1-5a34-48f6-9f72-1e548f563b20"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.676793 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ff8p\" (UniqueName: \"kubernetes.io/projected/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-kube-api-access-9ff8p\") pod \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.676909 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-serving-cert\") pod \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.677181 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwz7w\" (UniqueName: \"kubernetes.io/projected/eff08ce1-5a34-48f6-9f72-1e548f563b20-kube-api-access-gwz7w\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.677195 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eff08ce1-5a34-48f6-9f72-1e548f563b20-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.677228 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.677245 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.677256 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eff08ce1-5a34-48f6-9f72-1e548f563b20-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.683475 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-kube-api-access-9ff8p" (OuterVolumeSpecName: "kube-api-access-9ff8p") pod "526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" (UID: "526aa2e6-ed7c-4a98-97c6-c9bda284b9f2"). InnerVolumeSpecName "kube-api-access-9ff8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.686111 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" (UID: "526aa2e6-ed7c-4a98-97c6-c9bda284b9f2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.778107 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-client-ca\") pod \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.778192 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-config\") pod \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\" (UID: \"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2\") " Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.778341 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ff8p\" (UniqueName: \"kubernetes.io/projected/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-kube-api-access-9ff8p\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.778357 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.778739 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-client-ca" (OuterVolumeSpecName: "client-ca") pod "526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" (UID: "526aa2e6-ed7c-4a98-97c6-c9bda284b9f2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.778872 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-config" (OuterVolumeSpecName: "config") pod "526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" (UID: "526aa2e6-ed7c-4a98-97c6-c9bda284b9f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.879665 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:24 crc kubenswrapper[4715]: I1210 09:39:24.880205 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.222165 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" event={"ID":"eff08ce1-5a34-48f6-9f72-1e548f563b20","Type":"ContainerDied","Data":"6d2cfb45fc605138a23086541ff36c0a433e941582f5b67805ac0113ccc99b6d"} Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.222194 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-j2l56" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.222225 4715 scope.go:117] "RemoveContainer" containerID="b5aad762bd37a1e7dc07b3550cd4a1bc33d28a93d1f23cffb1a73b6f40403908" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.224051 4715 generic.go:334] "Generic (PLEG): container finished" podID="526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" containerID="2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62" exitCode=0 Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.224102 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" event={"ID":"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2","Type":"ContainerDied","Data":"2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62"} Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.224119 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" event={"ID":"526aa2e6-ed7c-4a98-97c6-c9bda284b9f2","Type":"ContainerDied","Data":"5218c4f490680fe96494b692ba0c35823a07ba407a69d9b2fc95472f797e9589"} Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.224126 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.230690 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjntt" event={"ID":"4e744f3a-62da-4f4e-a7ae-5fec5c47873f","Type":"ContainerStarted","Data":"f47200b14f28e4f2e7233a4d688bab7094c18c62ae3b1856ebb1c3b3ac8c1254"} Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.251788 4715 scope.go:117] "RemoveContainer" containerID="2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.265527 4715 scope.go:117] "RemoveContainer" containerID="2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62" Dec 10 09:39:25 crc kubenswrapper[4715]: E1210 09:39:25.265853 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62\": container with ID starting with 2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62 not found: ID does not exist" containerID="2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.265879 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62"} err="failed to get container status \"2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62\": rpc error: code = NotFound desc = could not find container \"2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62\": container with ID starting with 2c2bed295dbdff1e320997b173b56f14b4dfd7336c3c3396ae60796b9fcb5b62 not found: ID does not exist" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.274235 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-j2l56"] Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.277192 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-j2l56"] Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.286121 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl"] Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.289893 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-z96cl"] Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.612228 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" path="/var/lib/kubelet/pods/526aa2e6-ed7c-4a98-97c6-c9bda284b9f2/volumes" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.612940 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eff08ce1-5a34-48f6-9f72-1e548f563b20" path="/var/lib/kubelet/pods/eff08ce1-5a34-48f6-9f72-1e548f563b20/volumes" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.634532 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc"] Dec 10 09:39:25 crc kubenswrapper[4715]: E1210 09:39:25.634811 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" containerName="route-controller-manager" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.634831 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" containerName="route-controller-manager" Dec 10 09:39:25 crc kubenswrapper[4715]: E1210 09:39:25.634857 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff08ce1-5a34-48f6-9f72-1e548f563b20" containerName="controller-manager" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.634865 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff08ce1-5a34-48f6-9f72-1e548f563b20" containerName="controller-manager" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.634982 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff08ce1-5a34-48f6-9f72-1e548f563b20" containerName="controller-manager" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.634998 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="526aa2e6-ed7c-4a98-97c6-c9bda284b9f2" containerName="route-controller-manager" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.635458 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.638381 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.638836 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.639106 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.639110 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.641593 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.642083 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.643010 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-56c799745b-26f9q"] Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.643694 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.647070 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.647105 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.647781 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.648082 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.649464 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc"] Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.650844 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.653521 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56c799745b-26f9q"] Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.654174 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.656193 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.791426 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-client-ca\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.791763 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpwns\" (UniqueName: \"kubernetes.io/projected/63463b81-3a30-42c5-a583-0da3d25a3ee4-kube-api-access-lpwns\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.791868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-serving-cert\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.792017 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-client-ca\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.792114 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mssfb\" (UniqueName: \"kubernetes.io/projected/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-kube-api-access-mssfb\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.792203 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-proxy-ca-bundles\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.792313 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-config\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.792390 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63463b81-3a30-42c5-a583-0da3d25a3ee4-serving-cert\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.792497 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-config\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.895549 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpwns\" (UniqueName: \"kubernetes.io/projected/63463b81-3a30-42c5-a583-0da3d25a3ee4-kube-api-access-lpwns\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.895613 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-serving-cert\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.895679 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-client-ca\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.895714 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mssfb\" (UniqueName: \"kubernetes.io/projected/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-kube-api-access-mssfb\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.895747 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-proxy-ca-bundles\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.895794 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-config\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.895827 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63463b81-3a30-42c5-a583-0da3d25a3ee4-serving-cert\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.895882 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-config\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.895944 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-client-ca\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.898322 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-config\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.898774 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-proxy-ca-bundles\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.899728 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-config\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.900218 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-client-ca\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.901956 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63463b81-3a30-42c5-a583-0da3d25a3ee4-serving-cert\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.902617 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-client-ca\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.917085 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mssfb\" (UniqueName: \"kubernetes.io/projected/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-kube-api-access-mssfb\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.917386 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-serving-cert\") pod \"route-controller-manager-546d9447f4-7nmvc\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.919677 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpwns\" (UniqueName: \"kubernetes.io/projected/63463b81-3a30-42c5-a583-0da3d25a3ee4-kube-api-access-lpwns\") pod \"controller-manager-56c799745b-26f9q\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.951483 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:25 crc kubenswrapper[4715]: I1210 09:39:25.961425 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:26 crc kubenswrapper[4715]: I1210 09:39:26.157370 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56c799745b-26f9q"] Dec 10 09:39:26 crc kubenswrapper[4715]: W1210 09:39:26.158206 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63463b81_3a30_42c5_a583_0da3d25a3ee4.slice/crio-d2a3b6c6d30ceb49e134d0b875d23aea5179ec89bd43762d30ea9336c622e464 WatchSource:0}: Error finding container d2a3b6c6d30ceb49e134d0b875d23aea5179ec89bd43762d30ea9336c622e464: Status 404 returned error can't find the container with id d2a3b6c6d30ceb49e134d0b875d23aea5179ec89bd43762d30ea9336c622e464 Dec 10 09:39:26 crc kubenswrapper[4715]: I1210 09:39:26.191727 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc"] Dec 10 09:39:26 crc kubenswrapper[4715]: I1210 09:39:26.238441 4715 generic.go:334] "Generic (PLEG): container finished" podID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerID="f47200b14f28e4f2e7233a4d688bab7094c18c62ae3b1856ebb1c3b3ac8c1254" exitCode=0 Dec 10 09:39:26 crc kubenswrapper[4715]: I1210 09:39:26.238514 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjntt" event={"ID":"4e744f3a-62da-4f4e-a7ae-5fec5c47873f","Type":"ContainerDied","Data":"f47200b14f28e4f2e7233a4d688bab7094c18c62ae3b1856ebb1c3b3ac8c1254"} Dec 10 09:39:26 crc kubenswrapper[4715]: I1210 09:39:26.241908 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" event={"ID":"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a","Type":"ContainerStarted","Data":"152eaf9a6de6772c5144eb1bd92413c37d5c815030828a16333c7a35f6126dbb"} Dec 10 09:39:26 crc kubenswrapper[4715]: I1210 09:39:26.245258 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" event={"ID":"63463b81-3a30-42c5-a583-0da3d25a3ee4","Type":"ContainerStarted","Data":"d2a3b6c6d30ceb49e134d0b875d23aea5179ec89bd43762d30ea9336c622e464"} Dec 10 09:39:27 crc kubenswrapper[4715]: I1210 09:39:27.253402 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" event={"ID":"63463b81-3a30-42c5-a583-0da3d25a3ee4","Type":"ContainerStarted","Data":"c75e9166f24e3162dd14a101a1f08aacc24314ecb69ebdd9cffaa21987368c3f"} Dec 10 09:39:27 crc kubenswrapper[4715]: I1210 09:39:27.253941 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:27 crc kubenswrapper[4715]: I1210 09:39:27.255805 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" event={"ID":"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a","Type":"ContainerStarted","Data":"4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1"} Dec 10 09:39:27 crc kubenswrapper[4715]: I1210 09:39:27.256425 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:27 crc kubenswrapper[4715]: I1210 09:39:27.262056 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:27 crc kubenswrapper[4715]: I1210 09:39:27.263448 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:27 crc kubenswrapper[4715]: I1210 09:39:27.275582 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" podStartSLOduration=3.275556945 podStartE2EDuration="3.275556945s" podCreationTimestamp="2025-12-10 09:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:39:27.271040952 +0000 UTC m=+330.014587233" watchObservedRunningTime="2025-12-10 09:39:27.275556945 +0000 UTC m=+330.019103236" Dec 10 09:39:27 crc kubenswrapper[4715]: I1210 09:39:27.310235 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" podStartSLOduration=3.310214814 podStartE2EDuration="3.310214814s" podCreationTimestamp="2025-12-10 09:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:39:27.308029379 +0000 UTC m=+330.051575630" watchObservedRunningTime="2025-12-10 09:39:27.310214814 +0000 UTC m=+330.053761065" Dec 10 09:39:28 crc kubenswrapper[4715]: I1210 09:39:28.271705 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjntt" event={"ID":"4e744f3a-62da-4f4e-a7ae-5fec5c47873f","Type":"ContainerStarted","Data":"9d6415e023064b246fc7943f6671a6d94a80084d3abaa00c021a524954530a79"} Dec 10 09:39:28 crc kubenswrapper[4715]: I1210 09:39:28.290003 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tjntt" podStartSLOduration=3.2952806629999998 podStartE2EDuration="6.28998637s" podCreationTimestamp="2025-12-10 09:39:22 +0000 UTC" firstStartedPulling="2025-12-10 09:39:24.197357626 +0000 UTC m=+326.940903877" lastFinishedPulling="2025-12-10 09:39:27.192063333 +0000 UTC m=+329.935609584" observedRunningTime="2025-12-10 09:39:28.287423806 +0000 UTC m=+331.030970057" watchObservedRunningTime="2025-12-10 09:39:28.28998637 +0000 UTC m=+331.033532621" Dec 10 09:39:30 crc kubenswrapper[4715]: I1210 09:39:30.424178 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:30 crc kubenswrapper[4715]: I1210 09:39:30.424268 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:30 crc kubenswrapper[4715]: I1210 09:39:30.486396 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:30 crc kubenswrapper[4715]: I1210 09:39:30.639381 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:30 crc kubenswrapper[4715]: I1210 09:39:30.639463 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:30 crc kubenswrapper[4715]: I1210 09:39:30.677999 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:31 crc kubenswrapper[4715]: I1210 09:39:31.327384 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 09:39:31 crc kubenswrapper[4715]: I1210 09:39:31.331520 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nv8h7" Dec 10 09:39:33 crc kubenswrapper[4715]: I1210 09:39:33.215771 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:33 crc kubenswrapper[4715]: I1210 09:39:33.216830 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:33 crc kubenswrapper[4715]: I1210 09:39:33.256200 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:33 crc kubenswrapper[4715]: I1210 09:39:33.339847 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tjntt" Dec 10 09:39:36 crc kubenswrapper[4715]: I1210 09:39:36.326302 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56c799745b-26f9q"] Dec 10 09:39:36 crc kubenswrapper[4715]: I1210 09:39:36.326800 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" podUID="63463b81-3a30-42c5-a583-0da3d25a3ee4" containerName="controller-manager" containerID="cri-o://c75e9166f24e3162dd14a101a1f08aacc24314ecb69ebdd9cffaa21987368c3f" gracePeriod=30 Dec 10 09:39:36 crc kubenswrapper[4715]: I1210 09:39:36.344790 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc"] Dec 10 09:39:36 crc kubenswrapper[4715]: I1210 09:39:36.345032 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" podUID="cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a" containerName="route-controller-manager" containerID="cri-o://4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1" gracePeriod=30 Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.326488 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb8cn" event={"ID":"63a49347-3b45-4bd9-a158-6deb79a28d2d","Type":"ContainerStarted","Data":"ca48259ca1b22f0a9a18754b000af60f73f563c5c24ead63ddf19af59a5079ef"} Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.326831 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.328123 4715 generic.go:334] "Generic (PLEG): container finished" podID="63463b81-3a30-42c5-a583-0da3d25a3ee4" containerID="c75e9166f24e3162dd14a101a1f08aacc24314ecb69ebdd9cffaa21987368c3f" exitCode=0 Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.328232 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" event={"ID":"63463b81-3a30-42c5-a583-0da3d25a3ee4","Type":"ContainerDied","Data":"c75e9166f24e3162dd14a101a1f08aacc24314ecb69ebdd9cffaa21987368c3f"} Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.329322 4715 generic.go:334] "Generic (PLEG): container finished" podID="cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a" containerID="4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1" exitCode=0 Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.329341 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" event={"ID":"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a","Type":"ContainerDied","Data":"4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1"} Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.329355 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" event={"ID":"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a","Type":"ContainerDied","Data":"152eaf9a6de6772c5144eb1bd92413c37d5c815030828a16333c7a35f6126dbb"} Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.329371 4715 scope.go:117] "RemoveContainer" containerID="4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.329462 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.355475 4715 scope.go:117] "RemoveContainer" containerID="4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1" Dec 10 09:39:37 crc kubenswrapper[4715]: E1210 09:39:37.356468 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1\": container with ID starting with 4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1 not found: ID does not exist" containerID="4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.356510 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1"} err="failed to get container status \"4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1\": rpc error: code = NotFound desc = could not find container \"4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1\": container with ID starting with 4ed37e004173e1f098a8b2939fd4861c9dd863e9d2bedcde76c2a7bd19ee43c1 not found: ID does not exist" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.411326 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.445180 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-serving-cert\") pod \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.445230 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-client-ca\") pod \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.445272 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-client-ca\") pod \"63463b81-3a30-42c5-a583-0da3d25a3ee4\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.445313 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-config\") pod \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.445427 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-config\") pod \"63463b81-3a30-42c5-a583-0da3d25a3ee4\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.445472 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpwns\" (UniqueName: \"kubernetes.io/projected/63463b81-3a30-42c5-a583-0da3d25a3ee4-kube-api-access-lpwns\") pod \"63463b81-3a30-42c5-a583-0da3d25a3ee4\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.445504 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-proxy-ca-bundles\") pod \"63463b81-3a30-42c5-a583-0da3d25a3ee4\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.445531 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mssfb\" (UniqueName: \"kubernetes.io/projected/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-kube-api-access-mssfb\") pod \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\" (UID: \"cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a\") " Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.445555 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63463b81-3a30-42c5-a583-0da3d25a3ee4-serving-cert\") pod \"63463b81-3a30-42c5-a583-0da3d25a3ee4\" (UID: \"63463b81-3a30-42c5-a583-0da3d25a3ee4\") " Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.446009 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-client-ca" (OuterVolumeSpecName: "client-ca") pod "cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a" (UID: "cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.446056 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-config" (OuterVolumeSpecName: "config") pod "cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a" (UID: "cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.446423 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-client-ca" (OuterVolumeSpecName: "client-ca") pod "63463b81-3a30-42c5-a583-0da3d25a3ee4" (UID: "63463b81-3a30-42c5-a583-0da3d25a3ee4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.446432 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "63463b81-3a30-42c5-a583-0da3d25a3ee4" (UID: "63463b81-3a30-42c5-a583-0da3d25a3ee4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.447127 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-config" (OuterVolumeSpecName: "config") pod "63463b81-3a30-42c5-a583-0da3d25a3ee4" (UID: "63463b81-3a30-42c5-a583-0da3d25a3ee4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.450224 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63463b81-3a30-42c5-a583-0da3d25a3ee4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "63463b81-3a30-42c5-a583-0da3d25a3ee4" (UID: "63463b81-3a30-42c5-a583-0da3d25a3ee4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.450245 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63463b81-3a30-42c5-a583-0da3d25a3ee4-kube-api-access-lpwns" (OuterVolumeSpecName: "kube-api-access-lpwns") pod "63463b81-3a30-42c5-a583-0da3d25a3ee4" (UID: "63463b81-3a30-42c5-a583-0da3d25a3ee4"). InnerVolumeSpecName "kube-api-access-lpwns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.450520 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-kube-api-access-mssfb" (OuterVolumeSpecName: "kube-api-access-mssfb") pod "cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a" (UID: "cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a"). InnerVolumeSpecName "kube-api-access-mssfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.451383 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a" (UID: "cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.551228 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.551269 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mssfb\" (UniqueName: \"kubernetes.io/projected/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-kube-api-access-mssfb\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.551287 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63463b81-3a30-42c5-a583-0da3d25a3ee4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.551303 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.551323 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.551337 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.551349 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.551363 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63463b81-3a30-42c5-a583-0da3d25a3ee4-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.551381 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpwns\" (UniqueName: \"kubernetes.io/projected/63463b81-3a30-42c5-a583-0da3d25a3ee4-kube-api-access-lpwns\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.649099 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j"] Dec 10 09:39:37 crc kubenswrapper[4715]: E1210 09:39:37.649369 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63463b81-3a30-42c5-a583-0da3d25a3ee4" containerName="controller-manager" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.649385 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="63463b81-3a30-42c5-a583-0da3d25a3ee4" containerName="controller-manager" Dec 10 09:39:37 crc kubenswrapper[4715]: E1210 09:39:37.649403 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a" containerName="route-controller-manager" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.649413 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a" containerName="route-controller-manager" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.649533 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a" containerName="route-controller-manager" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.649551 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="63463b81-3a30-42c5-a583-0da3d25a3ee4" containerName="controller-manager" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.650053 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.652510 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-client-ca\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.652598 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-config\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.652631 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2l6x\" (UniqueName: \"kubernetes.io/projected/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-kube-api-access-w2l6x\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.652661 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-serving-cert\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.655065 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.655248 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.655515 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.655751 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.655755 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.658513 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j"] Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.659209 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.716754 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc"] Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.716841 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-546d9447f4-7nmvc"] Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.753619 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-config\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.753665 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2l6x\" (UniqueName: \"kubernetes.io/projected/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-kube-api-access-w2l6x\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.753699 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-serving-cert\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.753763 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-client-ca\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.754986 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-client-ca\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.755046 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-config\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.758342 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-serving-cert\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.771677 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2l6x\" (UniqueName: \"kubernetes.io/projected/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-kube-api-access-w2l6x\") pod \"route-controller-manager-7fdbcb8884-qtr6j\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.799491 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:37 crc kubenswrapper[4715]: I1210 09:39:37.982373 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j"] Dec 10 09:39:37 crc kubenswrapper[4715]: W1210 09:39:37.989978 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7d5b7f7_4b79_49dd_ab66_ca66b32bc387.slice/crio-27f6012f60fa7a12d0d15a908d754941c63fb206cfafb629ee5f0fd9b0ba73b7 WatchSource:0}: Error finding container 27f6012f60fa7a12d0d15a908d754941c63fb206cfafb629ee5f0fd9b0ba73b7: Status 404 returned error can't find the container with id 27f6012f60fa7a12d0d15a908d754941c63fb206cfafb629ee5f0fd9b0ba73b7 Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.337754 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.337778 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56c799745b-26f9q" event={"ID":"63463b81-3a30-42c5-a583-0da3d25a3ee4","Type":"ContainerDied","Data":"d2a3b6c6d30ceb49e134d0b875d23aea5179ec89bd43762d30ea9336c622e464"} Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.338317 4715 scope.go:117] "RemoveContainer" containerID="c75e9166f24e3162dd14a101a1f08aacc24314ecb69ebdd9cffaa21987368c3f" Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.342315 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" event={"ID":"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387","Type":"ContainerStarted","Data":"30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6"} Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.342371 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" event={"ID":"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387","Type":"ContainerStarted","Data":"27f6012f60fa7a12d0d15a908d754941c63fb206cfafb629ee5f0fd9b0ba73b7"} Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.342656 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.347729 4715 generic.go:334] "Generic (PLEG): container finished" podID="63a49347-3b45-4bd9-a158-6deb79a28d2d" containerID="ca48259ca1b22f0a9a18754b000af60f73f563c5c24ead63ddf19af59a5079ef" exitCode=0 Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.347767 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb8cn" event={"ID":"63a49347-3b45-4bd9-a158-6deb79a28d2d","Type":"ContainerDied","Data":"ca48259ca1b22f0a9a18754b000af60f73f563c5c24ead63ddf19af59a5079ef"} Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.382456 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" podStartSLOduration=2.382437577 podStartE2EDuration="2.382437577s" podCreationTimestamp="2025-12-10 09:39:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:39:38.366504188 +0000 UTC m=+341.110050529" watchObservedRunningTime="2025-12-10 09:39:38.382437577 +0000 UTC m=+341.125983828" Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.383941 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56c799745b-26f9q"] Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.387120 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-56c799745b-26f9q"] Dec 10 09:39:38 crc kubenswrapper[4715]: I1210 09:39:38.514164 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.356742 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb8cn" event={"ID":"63a49347-3b45-4bd9-a158-6deb79a28d2d","Type":"ContainerStarted","Data":"2b6a89babca52de00ed30b635dab8f168a28915910baebb8f0e9bbbf7d177586"} Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.376001 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hb8cn" podStartSLOduration=2.732216382 podStartE2EDuration="17.375985349s" podCreationTimestamp="2025-12-10 09:39:22 +0000 UTC" firstStartedPulling="2025-12-10 09:39:24.189317075 +0000 UTC m=+326.932863316" lastFinishedPulling="2025-12-10 09:39:38.833086032 +0000 UTC m=+341.576632283" observedRunningTime="2025-12-10 09:39:39.375750513 +0000 UTC m=+342.119296784" watchObservedRunningTime="2025-12-10 09:39:39.375985349 +0000 UTC m=+342.119531600" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.611741 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63463b81-3a30-42c5-a583-0da3d25a3ee4" path="/var/lib/kubelet/pods/63463b81-3a30-42c5-a583-0da3d25a3ee4/volumes" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.612406 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a" path="/var/lib/kubelet/pods/cc8c9eff-bc6b-4f7d-a938-79ff5db8fb0a/volumes" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.643798 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-76c96b674b-lbh92"] Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.644651 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.648492 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.648715 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.649127 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.649596 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.649823 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.649955 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.658338 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76c96b674b-lbh92"] Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.658341 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.700880 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-serving-cert\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.701001 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d7cn\" (UniqueName: \"kubernetes.io/projected/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-kube-api-access-7d7cn\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.701417 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-proxy-ca-bundles\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.701450 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-config\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.701497 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-client-ca\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.795523 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j"] Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.802637 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d7cn\" (UniqueName: \"kubernetes.io/projected/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-kube-api-access-7d7cn\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.802694 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-proxy-ca-bundles\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.802716 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-config\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.802742 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-client-ca\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.802809 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-serving-cert\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.804793 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-client-ca\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.805253 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-config\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.805351 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-proxy-ca-bundles\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.807394 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-serving-cert\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.833616 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d7cn\" (UniqueName: \"kubernetes.io/projected/c5df4e17-4f40-44a8-bcf1-b59b225c52f9-kube-api-access-7d7cn\") pod \"controller-manager-76c96b674b-lbh92\" (UID: \"c5df4e17-4f40-44a8-bcf1-b59b225c52f9\") " pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:39 crc kubenswrapper[4715]: I1210 09:39:39.962338 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:40 crc kubenswrapper[4715]: I1210 09:39:40.399799 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76c96b674b-lbh92"] Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.369511 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" event={"ID":"c5df4e17-4f40-44a8-bcf1-b59b225c52f9","Type":"ContainerStarted","Data":"a2819684b141bae387032297df985ce4400f9b898c036409bd19a7454446b28d"} Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.369863 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" event={"ID":"c5df4e17-4f40-44a8-bcf1-b59b225c52f9","Type":"ContainerStarted","Data":"60d9d24a3423fd31765c3f005d0155352716f26db74644735ba83d292b70d5aa"} Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.369695 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" podUID="c7d5b7f7-4b79-49dd-ab66-ca66b32bc387" containerName="route-controller-manager" containerID="cri-o://30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6" gracePeriod=30 Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.399942 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" podStartSLOduration=5.399901424 podStartE2EDuration="5.399901424s" podCreationTimestamp="2025-12-10 09:39:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:39:41.397988886 +0000 UTC m=+344.141535137" watchObservedRunningTime="2025-12-10 09:39:41.399901424 +0000 UTC m=+344.143447675" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.725896 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.734959 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2l6x\" (UniqueName: \"kubernetes.io/projected/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-kube-api-access-w2l6x\") pod \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.734997 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-config\") pod \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.735020 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-serving-cert\") pod \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.735056 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-client-ca\") pod \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\" (UID: \"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387\") " Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.735694 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-client-ca" (OuterVolumeSpecName: "client-ca") pod "c7d5b7f7-4b79-49dd-ab66-ca66b32bc387" (UID: "c7d5b7f7-4b79-49dd-ab66-ca66b32bc387"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.735776 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-config" (OuterVolumeSpecName: "config") pod "c7d5b7f7-4b79-49dd-ab66-ca66b32bc387" (UID: "c7d5b7f7-4b79-49dd-ab66-ca66b32bc387"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.741216 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c7d5b7f7-4b79-49dd-ab66-ca66b32bc387" (UID: "c7d5b7f7-4b79-49dd-ab66-ca66b32bc387"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.742365 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-kube-api-access-w2l6x" (OuterVolumeSpecName: "kube-api-access-w2l6x") pod "c7d5b7f7-4b79-49dd-ab66-ca66b32bc387" (UID: "c7d5b7f7-4b79-49dd-ab66-ca66b32bc387"). InnerVolumeSpecName "kube-api-access-w2l6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.751943 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7"] Dec 10 09:39:41 crc kubenswrapper[4715]: E1210 09:39:41.752195 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d5b7f7-4b79-49dd-ab66-ca66b32bc387" containerName="route-controller-manager" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.752216 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d5b7f7-4b79-49dd-ab66-ca66b32bc387" containerName="route-controller-manager" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.752312 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7d5b7f7-4b79-49dd-ab66-ca66b32bc387" containerName="route-controller-manager" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.752668 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.762868 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7"] Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.836270 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2l6x\" (UniqueName: \"kubernetes.io/projected/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-kube-api-access-w2l6x\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.836305 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.836315 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.836324 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.937797 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89a3eafa-8f84-4886-8464-a48d1b97c835-serving-cert\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.937895 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkvcw\" (UniqueName: \"kubernetes.io/projected/89a3eafa-8f84-4886-8464-a48d1b97c835-kube-api-access-lkvcw\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.937953 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-config\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:41 crc kubenswrapper[4715]: I1210 09:39:41.937979 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-client-ca\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.038696 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89a3eafa-8f84-4886-8464-a48d1b97c835-serving-cert\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.038776 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkvcw\" (UniqueName: \"kubernetes.io/projected/89a3eafa-8f84-4886-8464-a48d1b97c835-kube-api-access-lkvcw\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.038802 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-config\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.038821 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-client-ca\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.039871 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-client-ca\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.040113 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-config\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.042903 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89a3eafa-8f84-4886-8464-a48d1b97c835-serving-cert\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.063654 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkvcw\" (UniqueName: \"kubernetes.io/projected/89a3eafa-8f84-4886-8464-a48d1b97c835-kube-api-access-lkvcw\") pod \"route-controller-manager-76d65966c5-t2lc7\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.069719 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.380412 4715 generic.go:334] "Generic (PLEG): container finished" podID="c7d5b7f7-4b79-49dd-ab66-ca66b32bc387" containerID="30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6" exitCode=0 Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.380471 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.380470 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" event={"ID":"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387","Type":"ContainerDied","Data":"30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6"} Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.380621 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j" event={"ID":"c7d5b7f7-4b79-49dd-ab66-ca66b32bc387","Type":"ContainerDied","Data":"27f6012f60fa7a12d0d15a908d754941c63fb206cfafb629ee5f0fd9b0ba73b7"} Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.380655 4715 scope.go:117] "RemoveContainer" containerID="30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.380996 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.387685 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-76c96b674b-lbh92" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.399260 4715 scope.go:117] "RemoveContainer" containerID="30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6" Dec 10 09:39:42 crc kubenswrapper[4715]: E1210 09:39:42.399822 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6\": container with ID starting with 30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6 not found: ID does not exist" containerID="30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.399864 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6"} err="failed to get container status \"30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6\": rpc error: code = NotFound desc = could not find container \"30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6\": container with ID starting with 30d66b0ef5374bcc6ae062cc7438ad31f9e0f0347024c4b04fa06474fa94cec6 not found: ID does not exist" Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.439330 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j"] Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.442942 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fdbcb8884-qtr6j"] Dec 10 09:39:42 crc kubenswrapper[4715]: I1210 09:39:42.463296 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7"] Dec 10 09:39:42 crc kubenswrapper[4715]: W1210 09:39:42.470473 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89a3eafa_8f84_4886_8464_a48d1b97c835.slice/crio-714f56dbcedf65406924ab3b18c788c377ca0ed807003299cbfe1cd429b926c6 WatchSource:0}: Error finding container 714f56dbcedf65406924ab3b18c788c377ca0ed807003299cbfe1cd429b926c6: Status 404 returned error can't find the container with id 714f56dbcedf65406924ab3b18c788c377ca0ed807003299cbfe1cd429b926c6 Dec 10 09:39:43 crc kubenswrapper[4715]: I1210 09:39:43.024375 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:43 crc kubenswrapper[4715]: I1210 09:39:43.025234 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:43 crc kubenswrapper[4715]: I1210 09:39:43.388586 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" event={"ID":"89a3eafa-8f84-4886-8464-a48d1b97c835","Type":"ContainerStarted","Data":"8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940"} Dec 10 09:39:43 crc kubenswrapper[4715]: I1210 09:39:43.388640 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" event={"ID":"89a3eafa-8f84-4886-8464-a48d1b97c835","Type":"ContainerStarted","Data":"714f56dbcedf65406924ab3b18c788c377ca0ed807003299cbfe1cd429b926c6"} Dec 10 09:39:43 crc kubenswrapper[4715]: I1210 09:39:43.388925 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:43 crc kubenswrapper[4715]: I1210 09:39:43.415216 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" podStartSLOduration=4.415201494 podStartE2EDuration="4.415201494s" podCreationTimestamp="2025-12-10 09:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:39:43.41380966 +0000 UTC m=+346.157355931" watchObservedRunningTime="2025-12-10 09:39:43.415201494 +0000 UTC m=+346.158747735" Dec 10 09:39:43 crc kubenswrapper[4715]: I1210 09:39:43.422362 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:43 crc kubenswrapper[4715]: I1210 09:39:43.612095 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7d5b7f7-4b79-49dd-ab66-ca66b32bc387" path="/var/lib/kubelet/pods/c7d5b7f7-4b79-49dd-ab66-ca66b32bc387/volumes" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.073340 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hb8cn" podUID="63a49347-3b45-4bd9-a158-6deb79a28d2d" containerName="registry-server" probeResult="failure" output=< Dec 10 09:39:44 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 10 09:39:44 crc kubenswrapper[4715]: > Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.118840 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" podUID="e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" containerName="oauth-openshift" containerID="cri-o://1554d8c21cf416b54f7d32cfe94d127aa45e890f2ca309621101f07a15b7bfe1" gracePeriod=15 Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.466232 4715 generic.go:334] "Generic (PLEG): container finished" podID="e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" containerID="1554d8c21cf416b54f7d32cfe94d127aa45e890f2ca309621101f07a15b7bfe1" exitCode=0 Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.466278 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" event={"ID":"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1","Type":"ContainerDied","Data":"1554d8c21cf416b54f7d32cfe94d127aa45e890f2ca309621101f07a15b7bfe1"} Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.555851 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7"] Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.612604 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.769531 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-cliconfig\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.769590 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-trusted-ca-bundle\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.769616 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-login\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.769637 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-router-certs\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.769666 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-provider-selection\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.769696 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-ocp-branding-template\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.769748 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r26qr\" (UniqueName: \"kubernetes.io/projected/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-kube-api-access-r26qr\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.770322 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.770559 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-idp-0-file-data\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.770599 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-service-ca\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.770651 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-error\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.770679 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-dir\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.770708 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-serving-cert\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.770759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-session\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.770765 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.770812 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-policies\") pod \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\" (UID: \"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1\") " Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.771039 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.771067 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.771099 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.771117 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.771663 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.779536 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.779605 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-kube-api-access-r26qr" (OuterVolumeSpecName: "kube-api-access-r26qr") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "kube-api-access-r26qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.782259 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.782464 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.794318 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.794610 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.798476 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.798700 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.798858 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" (UID: "e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.872308 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.872673 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.872752 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.872881 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.873007 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r26qr\" (UniqueName: \"kubernetes.io/projected/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-kube-api-access-r26qr\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.873079 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.873155 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.873253 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.873346 4715 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.873431 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.873511 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:44 crc kubenswrapper[4715]: I1210 09:39:44.873589 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.370011 4715 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-7qnsk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.370376 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" podUID="e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.473586 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" event={"ID":"e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1","Type":"ContainerDied","Data":"632646a0ad77fc64faaa1e572dcb023922c9e5055e6be858d5e8d2ddab66efde"} Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.473620 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7qnsk" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.473668 4715 scope.go:117] "RemoveContainer" containerID="1554d8c21cf416b54f7d32cfe94d127aa45e890f2ca309621101f07a15b7bfe1" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.505612 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7qnsk"] Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.508525 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7qnsk"] Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.614058 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" path="/var/lib/kubelet/pods/e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1/volumes" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.648623 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-586d5b9769-wkmtq"] Dec 10 09:39:45 crc kubenswrapper[4715]: E1210 09:39:45.648863 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" containerName="oauth-openshift" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.648875 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" containerName="oauth-openshift" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.649021 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7a77c53-aea3-40ab-8b5b-2c6c51b83ff1" containerName="oauth-openshift" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.649434 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.653836 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.654731 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.654765 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.655051 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.655155 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.655677 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.655995 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.656200 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.656218 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.656808 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.656938 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.657079 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.666207 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.670585 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.687699 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.705337 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-586d5b9769-wkmtq"] Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.788747 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.788981 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789020 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-dir\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789044 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789069 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-error\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789132 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-login\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789206 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-session\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789238 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78sm9\" (UniqueName: \"kubernetes.io/projected/0eb53283-1c1d-4a93-a160-7d5f591e85f1-kube-api-access-78sm9\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789308 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-policies\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789325 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-service-ca\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789344 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789362 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-router-certs\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789393 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.789454 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.891410 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.891534 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.891595 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.891644 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.891708 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-dir\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.891762 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.891802 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-error\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.891860 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-login\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.891975 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-session\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.892039 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78sm9\" (UniqueName: \"kubernetes.io/projected/0eb53283-1c1d-4a93-a160-7d5f591e85f1-kube-api-access-78sm9\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.892119 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-policies\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.892163 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-service-ca\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.892221 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.892295 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-router-certs\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.893361 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.894143 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-service-ca\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.894612 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.895020 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-dir\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.897716 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-policies\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.898508 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-router-certs\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.899339 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-login\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.899475 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-error\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.900602 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.901116 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.901598 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.903313 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.904461 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-session\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.918139 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78sm9\" (UniqueName: \"kubernetes.io/projected/0eb53283-1c1d-4a93-a160-7d5f591e85f1-kube-api-access-78sm9\") pod \"oauth-openshift-586d5b9769-wkmtq\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:45 crc kubenswrapper[4715]: I1210 09:39:45.966942 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.217482 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mpsfh"] Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.218673 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.234986 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mpsfh"] Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.375819 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-586d5b9769-wkmtq"] Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.399069 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a450373-4901-4c81-99a4-d32f4aa25a77-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.399105 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a450373-4901-4c81-99a4-d32f4aa25a77-trusted-ca\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.399128 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a450373-4901-4c81-99a4-d32f4aa25a77-bound-sa-token\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.399175 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a450373-4901-4c81-99a4-d32f4aa25a77-registry-tls\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.399207 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sckqb\" (UniqueName: \"kubernetes.io/projected/8a450373-4901-4c81-99a4-d32f4aa25a77-kube-api-access-sckqb\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.399266 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a450373-4901-4c81-99a4-d32f4aa25a77-registry-certificates\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.399306 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.399345 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a450373-4901-4c81-99a4-d32f4aa25a77-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.426476 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.486023 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" podUID="89a3eafa-8f84-4886-8464-a48d1b97c835" containerName="route-controller-manager" containerID="cri-o://8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940" gracePeriod=30 Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.486461 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" event={"ID":"0eb53283-1c1d-4a93-a160-7d5f591e85f1","Type":"ContainerStarted","Data":"1d33f938635f4fc32fc1bc51870c3d8a5a9769a69001daea9a0065766977015b"} Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.503185 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a450373-4901-4c81-99a4-d32f4aa25a77-registry-certificates\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.503284 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a450373-4901-4c81-99a4-d32f4aa25a77-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.503332 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a450373-4901-4c81-99a4-d32f4aa25a77-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.503364 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a450373-4901-4c81-99a4-d32f4aa25a77-trusted-ca\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.503390 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a450373-4901-4c81-99a4-d32f4aa25a77-bound-sa-token\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.503420 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a450373-4901-4c81-99a4-d32f4aa25a77-registry-tls\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.503464 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sckqb\" (UniqueName: \"kubernetes.io/projected/8a450373-4901-4c81-99a4-d32f4aa25a77-kube-api-access-sckqb\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.504264 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a450373-4901-4c81-99a4-d32f4aa25a77-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.504858 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a450373-4901-4c81-99a4-d32f4aa25a77-registry-certificates\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.505079 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a450373-4901-4c81-99a4-d32f4aa25a77-trusted-ca\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.510773 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a450373-4901-4c81-99a4-d32f4aa25a77-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.515449 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a450373-4901-4c81-99a4-d32f4aa25a77-registry-tls\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.520807 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a450373-4901-4c81-99a4-d32f4aa25a77-bound-sa-token\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.523952 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sckqb\" (UniqueName: \"kubernetes.io/projected/8a450373-4901-4c81-99a4-d32f4aa25a77-kube-api-access-sckqb\") pod \"image-registry-66df7c8f76-mpsfh\" (UID: \"8a450373-4901-4c81-99a4-d32f4aa25a77\") " pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.534997 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:46 crc kubenswrapper[4715]: I1210 09:39:46.958380 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mpsfh"] Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.443656 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.493281 4715 generic.go:334] "Generic (PLEG): container finished" podID="89a3eafa-8f84-4886-8464-a48d1b97c835" containerID="8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940" exitCode=0 Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.493349 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" event={"ID":"89a3eafa-8f84-4886-8464-a48d1b97c835","Type":"ContainerDied","Data":"8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940"} Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.493378 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" event={"ID":"89a3eafa-8f84-4886-8464-a48d1b97c835","Type":"ContainerDied","Data":"714f56dbcedf65406924ab3b18c788c377ca0ed807003299cbfe1cd429b926c6"} Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.493398 4715 scope.go:117] "RemoveContainer" containerID="8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.493507 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.496734 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" event={"ID":"0eb53283-1c1d-4a93-a160-7d5f591e85f1","Type":"ContainerStarted","Data":"e5116523910924fcf408694395296c9999a8164797cded274770cad22972bef5"} Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.496792 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.500002 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" event={"ID":"8a450373-4901-4c81-99a4-d32f4aa25a77","Type":"ContainerStarted","Data":"fe012c55ef89be785d0af7d7524223702e1b0d6c59f73054c415304a3d882f15"} Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.500038 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" event={"ID":"8a450373-4901-4c81-99a4-d32f4aa25a77","Type":"ContainerStarted","Data":"7a935b30175e88f7cdb89139bf87e677def6d5faf6ac8258da00d59c5c70f4a3"} Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.500174 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.509866 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.519861 4715 scope.go:117] "RemoveContainer" containerID="8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940" Dec 10 09:39:47 crc kubenswrapper[4715]: E1210 09:39:47.520407 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940\": container with ID starting with 8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940 not found: ID does not exist" containerID="8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.520460 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940"} err="failed to get container status \"8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940\": rpc error: code = NotFound desc = could not find container \"8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940\": container with ID starting with 8670c470050d15a1c3ebe64d5f60922dbb64371826f8d3471cb73169c18af940 not found: ID does not exist" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.533260 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" podStartSLOduration=28.533207834 podStartE2EDuration="28.533207834s" podCreationTimestamp="2025-12-10 09:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:39:47.527241424 +0000 UTC m=+350.270787675" watchObservedRunningTime="2025-12-10 09:39:47.533207834 +0000 UTC m=+350.276754085" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.551109 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" podStartSLOduration=1.551088832 podStartE2EDuration="1.551088832s" podCreationTimestamp="2025-12-10 09:39:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:39:47.55059244 +0000 UTC m=+350.294138691" watchObservedRunningTime="2025-12-10 09:39:47.551088832 +0000 UTC m=+350.294635083" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.621773 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkvcw\" (UniqueName: \"kubernetes.io/projected/89a3eafa-8f84-4886-8464-a48d1b97c835-kube-api-access-lkvcw\") pod \"89a3eafa-8f84-4886-8464-a48d1b97c835\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.622006 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-config\") pod \"89a3eafa-8f84-4886-8464-a48d1b97c835\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.622083 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-client-ca\") pod \"89a3eafa-8f84-4886-8464-a48d1b97c835\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.622118 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89a3eafa-8f84-4886-8464-a48d1b97c835-serving-cert\") pod \"89a3eafa-8f84-4886-8464-a48d1b97c835\" (UID: \"89a3eafa-8f84-4886-8464-a48d1b97c835\") " Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.622847 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-client-ca" (OuterVolumeSpecName: "client-ca") pod "89a3eafa-8f84-4886-8464-a48d1b97c835" (UID: "89a3eafa-8f84-4886-8464-a48d1b97c835"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.622964 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.622968 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-config" (OuterVolumeSpecName: "config") pod "89a3eafa-8f84-4886-8464-a48d1b97c835" (UID: "89a3eafa-8f84-4886-8464-a48d1b97c835"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.628646 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89a3eafa-8f84-4886-8464-a48d1b97c835-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "89a3eafa-8f84-4886-8464-a48d1b97c835" (UID: "89a3eafa-8f84-4886-8464-a48d1b97c835"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.628668 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89a3eafa-8f84-4886-8464-a48d1b97c835-kube-api-access-lkvcw" (OuterVolumeSpecName: "kube-api-access-lkvcw") pod "89a3eafa-8f84-4886-8464-a48d1b97c835" (UID: "89a3eafa-8f84-4886-8464-a48d1b97c835"). InnerVolumeSpecName "kube-api-access-lkvcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.652505 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg"] Dec 10 09:39:47 crc kubenswrapper[4715]: E1210 09:39:47.652742 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a3eafa-8f84-4886-8464-a48d1b97c835" containerName="route-controller-manager" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.652753 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a3eafa-8f84-4886-8464-a48d1b97c835" containerName="route-controller-manager" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.652849 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="89a3eafa-8f84-4886-8464-a48d1b97c835" containerName="route-controller-manager" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.653313 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.681334 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg"] Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.714002 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.714086 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.724458 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkvcw\" (UniqueName: \"kubernetes.io/projected/89a3eafa-8f84-4886-8464-a48d1b97c835-kube-api-access-lkvcw\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.724582 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89a3eafa-8f84-4886-8464-a48d1b97c835-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.724752 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89a3eafa-8f84-4886-8464-a48d1b97c835-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.826109 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/684386dc-fc37-4b8f-a332-480e9ec4733a-client-ca\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.826385 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/684386dc-fc37-4b8f-a332-480e9ec4733a-serving-cert\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.826605 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/684386dc-fc37-4b8f-a332-480e9ec4733a-config\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.826733 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzpwv\" (UniqueName: \"kubernetes.io/projected/684386dc-fc37-4b8f-a332-480e9ec4733a-kube-api-access-jzpwv\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.827871 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7"] Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.832485 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76d65966c5-t2lc7"] Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.928159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/684386dc-fc37-4b8f-a332-480e9ec4733a-config\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.928217 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzpwv\" (UniqueName: \"kubernetes.io/projected/684386dc-fc37-4b8f-a332-480e9ec4733a-kube-api-access-jzpwv\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.928255 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/684386dc-fc37-4b8f-a332-480e9ec4733a-client-ca\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.928269 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/684386dc-fc37-4b8f-a332-480e9ec4733a-serving-cert\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.929793 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/684386dc-fc37-4b8f-a332-480e9ec4733a-client-ca\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.930049 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/684386dc-fc37-4b8f-a332-480e9ec4733a-config\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.931578 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/684386dc-fc37-4b8f-a332-480e9ec4733a-serving-cert\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.945899 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzpwv\" (UniqueName: \"kubernetes.io/projected/684386dc-fc37-4b8f-a332-480e9ec4733a-kube-api-access-jzpwv\") pod \"route-controller-manager-7fdbcb8884-rhlwg\" (UID: \"684386dc-fc37-4b8f-a332-480e9ec4733a\") " pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:47 crc kubenswrapper[4715]: I1210 09:39:47.972233 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:48 crc kubenswrapper[4715]: I1210 09:39:48.397011 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg"] Dec 10 09:39:48 crc kubenswrapper[4715]: W1210 09:39:48.417176 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod684386dc_fc37_4b8f_a332_480e9ec4733a.slice/crio-dfd6da2e7d9ef6c7f76c4b810a775f86c13fc4733f51e7232b89fe34d70bafbb WatchSource:0}: Error finding container dfd6da2e7d9ef6c7f76c4b810a775f86c13fc4733f51e7232b89fe34d70bafbb: Status 404 returned error can't find the container with id dfd6da2e7d9ef6c7f76c4b810a775f86c13fc4733f51e7232b89fe34d70bafbb Dec 10 09:39:48 crc kubenswrapper[4715]: I1210 09:39:48.511095 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" event={"ID":"684386dc-fc37-4b8f-a332-480e9ec4733a","Type":"ContainerStarted","Data":"dfd6da2e7d9ef6c7f76c4b810a775f86c13fc4733f51e7232b89fe34d70bafbb"} Dec 10 09:39:49 crc kubenswrapper[4715]: I1210 09:39:49.611873 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89a3eafa-8f84-4886-8464-a48d1b97c835" path="/var/lib/kubelet/pods/89a3eafa-8f84-4886-8464-a48d1b97c835/volumes" Dec 10 09:39:50 crc kubenswrapper[4715]: I1210 09:39:50.542565 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" event={"ID":"684386dc-fc37-4b8f-a332-480e9ec4733a","Type":"ContainerStarted","Data":"2c1714734a6217d32e84ed44a177c22fef403dd980691d838be2173b3981aa9f"} Dec 10 09:39:50 crc kubenswrapper[4715]: I1210 09:39:50.542892 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:50 crc kubenswrapper[4715]: I1210 09:39:50.549141 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" Dec 10 09:39:50 crc kubenswrapper[4715]: I1210 09:39:50.562639 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7fdbcb8884-rhlwg" podStartSLOduration=6.5626203 podStartE2EDuration="6.5626203s" podCreationTimestamp="2025-12-10 09:39:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:39:50.559722088 +0000 UTC m=+353.303268349" watchObservedRunningTime="2025-12-10 09:39:50.5626203 +0000 UTC m=+353.306166551" Dec 10 09:39:53 crc kubenswrapper[4715]: I1210 09:39:53.068949 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:39:53 crc kubenswrapper[4715]: I1210 09:39:53.131109 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hb8cn" Dec 10 09:40:06 crc kubenswrapper[4715]: I1210 09:40:06.423695 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-586d5b9769-wkmtq"] Dec 10 09:40:06 crc kubenswrapper[4715]: I1210 09:40:06.542372 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-mpsfh" Dec 10 09:40:06 crc kubenswrapper[4715]: I1210 09:40:06.595849 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nllws"] Dec 10 09:40:17 crc kubenswrapper[4715]: I1210 09:40:17.717350 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:40:17 crc kubenswrapper[4715]: I1210 09:40:17.717995 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.462842 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" podUID="0eb53283-1c1d-4a93-a160-7d5f591e85f1" containerName="oauth-openshift" containerID="cri-o://e5116523910924fcf408694395296c9999a8164797cded274770cad22972bef5" gracePeriod=15 Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.635478 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" podUID="3dcd1e77-33f0-4a5b-95c5-7454b648db5c" containerName="registry" containerID="cri-o://d9d6ae89e361ab4c8ae2a25bba5e37f34eea0ca6fb9cc521576d331baddd5bb0" gracePeriod=30 Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.779708 4715 generic.go:334] "Generic (PLEG): container finished" podID="3dcd1e77-33f0-4a5b-95c5-7454b648db5c" containerID="d9d6ae89e361ab4c8ae2a25bba5e37f34eea0ca6fb9cc521576d331baddd5bb0" exitCode=0 Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.779886 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" event={"ID":"3dcd1e77-33f0-4a5b-95c5-7454b648db5c","Type":"ContainerDied","Data":"d9d6ae89e361ab4c8ae2a25bba5e37f34eea0ca6fb9cc521576d331baddd5bb0"} Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.781842 4715 generic.go:334] "Generic (PLEG): container finished" podID="0eb53283-1c1d-4a93-a160-7d5f591e85f1" containerID="e5116523910924fcf408694395296c9999a8164797cded274770cad22972bef5" exitCode=0 Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.781879 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" event={"ID":"0eb53283-1c1d-4a93-a160-7d5f591e85f1","Type":"ContainerDied","Data":"e5116523910924fcf408694395296c9999a8164797cded274770cad22972bef5"} Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.902156 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.930728 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-75756b7776-kd7tk"] Dec 10 09:40:31 crc kubenswrapper[4715]: E1210 09:40:31.930954 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eb53283-1c1d-4a93-a160-7d5f591e85f1" containerName="oauth-openshift" Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.930967 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eb53283-1c1d-4a93-a160-7d5f591e85f1" containerName="oauth-openshift" Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.931082 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eb53283-1c1d-4a93-a160-7d5f591e85f1" containerName="oauth-openshift" Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.931472 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:31 crc kubenswrapper[4715]: I1210 09:40:31.993211 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-75756b7776-kd7tk"] Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001480 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-service-ca\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001522 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-error\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001542 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-idp-0-file-data\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001578 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-session\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001615 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78sm9\" (UniqueName: \"kubernetes.io/projected/0eb53283-1c1d-4a93-a160-7d5f591e85f1-kube-api-access-78sm9\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001661 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-provider-selection\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001703 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-dir\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001726 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-login\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001750 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-router-certs\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001779 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-ocp-branding-template\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001801 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-cliconfig\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001848 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-trusted-ca-bundle\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001881 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-serving-cert\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.001900 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-policies\") pod \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\" (UID: \"0eb53283-1c1d-4a93-a160-7d5f591e85f1\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.002613 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.002709 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.004368 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.004901 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.004964 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.008171 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.008198 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.008420 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.009373 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.010107 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.010129 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.010183 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.014746 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.021470 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eb53283-1c1d-4a93-a160-7d5f591e85f1-kube-api-access-78sm9" (OuterVolumeSpecName: "kube-api-access-78sm9") pod "0eb53283-1c1d-4a93-a160-7d5f591e85f1" (UID: "0eb53283-1c1d-4a93-a160-7d5f591e85f1"). InnerVolumeSpecName "kube-api-access-78sm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.042401 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103075 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-template-error\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103156 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/533dcff0-8496-4fbd-9256-cd7a7999e1c5-audit-dir\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103186 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103204 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103262 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103287 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-session\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103302 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103505 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-service-ca\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103582 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-router-certs\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103666 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhzsl\" (UniqueName: \"kubernetes.io/projected/533dcff0-8496-4fbd-9256-cd7a7999e1c5-kube-api-access-mhzsl\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103766 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103828 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-audit-policies\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103882 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.103993 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-template-login\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104122 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104152 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104172 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104192 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104212 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78sm9\" (UniqueName: \"kubernetes.io/projected/0eb53283-1c1d-4a93-a160-7d5f591e85f1-kube-api-access-78sm9\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104231 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104251 4715 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104269 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104286 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104305 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104323 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104342 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104363 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0eb53283-1c1d-4a93-a160-7d5f591e85f1-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.104381 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0eb53283-1c1d-4a93-a160-7d5f591e85f1-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.205797 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.205889 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-ca-trust-extracted\") pod \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.205985 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-certificates\") pod \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206009 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-bound-sa-token\") pod \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206060 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-tls\") pod \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206096 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-trusted-ca\") pod \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206132 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-installation-pull-secrets\") pod \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206156 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhbx8\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-kube-api-access-zhbx8\") pod \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\" (UID: \"3dcd1e77-33f0-4a5b-95c5-7454b648db5c\") " Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206415 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206456 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206487 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206522 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-session\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206543 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206574 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-service-ca\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206595 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-router-certs\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206626 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhzsl\" (UniqueName: \"kubernetes.io/projected/533dcff0-8496-4fbd-9256-cd7a7999e1c5-kube-api-access-mhzsl\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206666 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206690 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-audit-policies\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206721 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206748 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-template-login\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206776 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-template-error\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206857 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/533dcff0-8496-4fbd-9256-cd7a7999e1c5-audit-dir\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.206993 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/533dcff0-8496-4fbd-9256-cd7a7999e1c5-audit-dir\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.208245 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.208526 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "3dcd1e77-33f0-4a5b-95c5-7454b648db5c" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.209703 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "3dcd1e77-33f0-4a5b-95c5-7454b648db5c" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.210846 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-audit-policies\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.211513 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.211668 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.213015 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "3dcd1e77-33f0-4a5b-95c5-7454b648db5c" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.213861 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-service-ca\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.215492 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "3dcd1e77-33f0-4a5b-95c5-7454b648db5c" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.216111 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.216352 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "3dcd1e77-33f0-4a5b-95c5-7454b648db5c" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.217015 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-template-login\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.217200 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-template-error\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.219477 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.219670 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-router-certs\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.219781 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-session\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.222008 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-kube-api-access-zhbx8" (OuterVolumeSpecName: "kube-api-access-zhbx8") pod "3dcd1e77-33f0-4a5b-95c5-7454b648db5c" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c"). InnerVolumeSpecName "kube-api-access-zhbx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.224596 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/533dcff0-8496-4fbd-9256-cd7a7999e1c5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.226168 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "3dcd1e77-33f0-4a5b-95c5-7454b648db5c" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.236287 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "3dcd1e77-33f0-4a5b-95c5-7454b648db5c" (UID: "3dcd1e77-33f0-4a5b-95c5-7454b648db5c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.244487 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhzsl\" (UniqueName: \"kubernetes.io/projected/533dcff0-8496-4fbd-9256-cd7a7999e1c5-kube-api-access-mhzsl\") pod \"oauth-openshift-75756b7776-kd7tk\" (UID: \"533dcff0-8496-4fbd-9256-cd7a7999e1c5\") " pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.293937 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.308524 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.308591 4715 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.308612 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhbx8\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-kube-api-access-zhbx8\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.308631 4715 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.308650 4715 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.308668 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.308685 4715 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3dcd1e77-33f0-4a5b-95c5-7454b648db5c-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.519431 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-75756b7776-kd7tk"] Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.790628 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" event={"ID":"3dcd1e77-33f0-4a5b-95c5-7454b648db5c","Type":"ContainerDied","Data":"16802a7acc9d1e69a52542c48f0314ace074a7d9a32658eae1ee4593ca3f4682"} Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.790675 4715 scope.go:117] "RemoveContainer" containerID="d9d6ae89e361ab4c8ae2a25bba5e37f34eea0ca6fb9cc521576d331baddd5bb0" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.790786 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nllws" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.793770 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" event={"ID":"533dcff0-8496-4fbd-9256-cd7a7999e1c5","Type":"ContainerStarted","Data":"0da7a2ca2e30534be51b366b7bf479ad6118df497ac2420a6e63eb2d2b0fd3b8"} Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.796737 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" event={"ID":"0eb53283-1c1d-4a93-a160-7d5f591e85f1","Type":"ContainerDied","Data":"1d33f938635f4fc32fc1bc51870c3d8a5a9769a69001daea9a0065766977015b"} Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.796796 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-586d5b9769-wkmtq" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.813662 4715 scope.go:117] "RemoveContainer" containerID="e5116523910924fcf408694395296c9999a8164797cded274770cad22972bef5" Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.822327 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nllws"] Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.825675 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nllws"] Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.838357 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-586d5b9769-wkmtq"] Dec 10 09:40:32 crc kubenswrapper[4715]: I1210 09:40:32.844099 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-586d5b9769-wkmtq"] Dec 10 09:40:33 crc kubenswrapper[4715]: I1210 09:40:33.616758 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0eb53283-1c1d-4a93-a160-7d5f591e85f1" path="/var/lib/kubelet/pods/0eb53283-1c1d-4a93-a160-7d5f591e85f1/volumes" Dec 10 09:40:33 crc kubenswrapper[4715]: I1210 09:40:33.617825 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dcd1e77-33f0-4a5b-95c5-7454b648db5c" path="/var/lib/kubelet/pods/3dcd1e77-33f0-4a5b-95c5-7454b648db5c/volumes" Dec 10 09:40:33 crc kubenswrapper[4715]: I1210 09:40:33.805624 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" event={"ID":"533dcff0-8496-4fbd-9256-cd7a7999e1c5","Type":"ContainerStarted","Data":"93085bd04f527a3c0b95ac2abcc7ca2f6f69da2556e2ac4e5b984e9c989da50e"} Dec 10 09:40:33 crc kubenswrapper[4715]: I1210 09:40:33.806411 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:33 crc kubenswrapper[4715]: I1210 09:40:33.811944 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" Dec 10 09:40:33 crc kubenswrapper[4715]: I1210 09:40:33.868244 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-75756b7776-kd7tk" podStartSLOduration=27.868222363 podStartE2EDuration="27.868222363s" podCreationTimestamp="2025-12-10 09:40:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:40:33.840377571 +0000 UTC m=+396.583923852" watchObservedRunningTime="2025-12-10 09:40:33.868222363 +0000 UTC m=+396.611768614" Dec 10 09:40:47 crc kubenswrapper[4715]: I1210 09:40:47.713848 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:40:47 crc kubenswrapper[4715]: I1210 09:40:47.714380 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:40:47 crc kubenswrapper[4715]: I1210 09:40:47.714423 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:40:47 crc kubenswrapper[4715]: I1210 09:40:47.715081 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f0e5a5883c2401a24c6955974c04dfd24314a74cbb02a997f9f8a68454c4153f"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 09:40:47 crc kubenswrapper[4715]: I1210 09:40:47.715154 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://f0e5a5883c2401a24c6955974c04dfd24314a74cbb02a997f9f8a68454c4153f" gracePeriod=600 Dec 10 09:40:48 crc kubenswrapper[4715]: I1210 09:40:48.888050 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="f0e5a5883c2401a24c6955974c04dfd24314a74cbb02a997f9f8a68454c4153f" exitCode=0 Dec 10 09:40:48 crc kubenswrapper[4715]: I1210 09:40:48.888155 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"f0e5a5883c2401a24c6955974c04dfd24314a74cbb02a997f9f8a68454c4153f"} Dec 10 09:40:48 crc kubenswrapper[4715]: I1210 09:40:48.888594 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"ed1ab5cf91267ff02613fe03163e2a0e1e6fe3ab62fe2bfb0b6ed4dcd711b4fe"} Dec 10 09:40:48 crc kubenswrapper[4715]: I1210 09:40:48.888625 4715 scope.go:117] "RemoveContainer" containerID="7c9b4128d0fe9a19910faea8525e4ceeda2acb916b880ed43af4e695a7b102f5" Dec 10 09:43:17 crc kubenswrapper[4715]: I1210 09:43:17.714747 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:43:17 crc kubenswrapper[4715]: I1210 09:43:17.715432 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:43:47 crc kubenswrapper[4715]: I1210 09:43:47.713827 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:43:47 crc kubenswrapper[4715]: I1210 09:43:47.714428 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.777755 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cxn4s"] Dec 10 09:44:13 crc kubenswrapper[4715]: E1210 09:44:13.778492 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dcd1e77-33f0-4a5b-95c5-7454b648db5c" containerName="registry" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.778507 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dcd1e77-33f0-4a5b-95c5-7454b648db5c" containerName="registry" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.778609 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dcd1e77-33f0-4a5b-95c5-7454b648db5c" containerName="registry" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.779005 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-cxn4s" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.786060 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.786087 4715 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-5246q" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.787384 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.792059 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-mmntd"] Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.792767 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-mmntd" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.794672 4715 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-q49vg" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.799073 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cxn4s"] Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.816194 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-mmntd"] Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.824544 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-vckw6"] Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.830484 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-vckw6" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.832772 4715 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-jjrkw" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.837870 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-vckw6"] Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.946654 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhchk\" (UniqueName: \"kubernetes.io/projected/189f8af2-11e6-4140-8104-fb4a66bce776-kube-api-access-jhchk\") pod \"cert-manager-cainjector-7f985d654d-cxn4s\" (UID: \"189f8af2-11e6-4140-8104-fb4a66bce776\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cxn4s" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.946948 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zl42\" (UniqueName: \"kubernetes.io/projected/7f4c1366-2e47-4899-a38f-55cbc26f0569-kube-api-access-5zl42\") pod \"cert-manager-webhook-5655c58dd6-vckw6\" (UID: \"7f4c1366-2e47-4899-a38f-55cbc26f0569\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-vckw6" Dec 10 09:44:13 crc kubenswrapper[4715]: I1210 09:44:13.947099 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf7cq\" (UniqueName: \"kubernetes.io/projected/7a38a68d-c3fe-40f5-be5d-4c75b919c4c4-kube-api-access-xf7cq\") pod \"cert-manager-5b446d88c5-mmntd\" (UID: \"7a38a68d-c3fe-40f5-be5d-4c75b919c4c4\") " pod="cert-manager/cert-manager-5b446d88c5-mmntd" Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.048107 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf7cq\" (UniqueName: \"kubernetes.io/projected/7a38a68d-c3fe-40f5-be5d-4c75b919c4c4-kube-api-access-xf7cq\") pod \"cert-manager-5b446d88c5-mmntd\" (UID: \"7a38a68d-c3fe-40f5-be5d-4c75b919c4c4\") " pod="cert-manager/cert-manager-5b446d88c5-mmntd" Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.048201 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhchk\" (UniqueName: \"kubernetes.io/projected/189f8af2-11e6-4140-8104-fb4a66bce776-kube-api-access-jhchk\") pod \"cert-manager-cainjector-7f985d654d-cxn4s\" (UID: \"189f8af2-11e6-4140-8104-fb4a66bce776\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cxn4s" Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.048258 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zl42\" (UniqueName: \"kubernetes.io/projected/7f4c1366-2e47-4899-a38f-55cbc26f0569-kube-api-access-5zl42\") pod \"cert-manager-webhook-5655c58dd6-vckw6\" (UID: \"7f4c1366-2e47-4899-a38f-55cbc26f0569\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-vckw6" Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.071630 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhchk\" (UniqueName: \"kubernetes.io/projected/189f8af2-11e6-4140-8104-fb4a66bce776-kube-api-access-jhchk\") pod \"cert-manager-cainjector-7f985d654d-cxn4s\" (UID: \"189f8af2-11e6-4140-8104-fb4a66bce776\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cxn4s" Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.072358 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zl42\" (UniqueName: \"kubernetes.io/projected/7f4c1366-2e47-4899-a38f-55cbc26f0569-kube-api-access-5zl42\") pod \"cert-manager-webhook-5655c58dd6-vckw6\" (UID: \"7f4c1366-2e47-4899-a38f-55cbc26f0569\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-vckw6" Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.073121 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf7cq\" (UniqueName: \"kubernetes.io/projected/7a38a68d-c3fe-40f5-be5d-4c75b919c4c4-kube-api-access-xf7cq\") pod \"cert-manager-5b446d88c5-mmntd\" (UID: \"7a38a68d-c3fe-40f5-be5d-4c75b919c4c4\") " pod="cert-manager/cert-manager-5b446d88c5-mmntd" Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.096075 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-cxn4s" Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.108940 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-mmntd" Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.144303 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-vckw6" Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.378304 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-mmntd"] Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.389944 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.641235 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cxn4s"] Dec 10 09:44:14 crc kubenswrapper[4715]: I1210 09:44:14.644151 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-vckw6"] Dec 10 09:44:14 crc kubenswrapper[4715]: W1210 09:44:14.649606 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4c1366_2e47_4899_a38f_55cbc26f0569.slice/crio-f46a58ed35edf7f8e88a8cbc53e692f6b9c389d8006a9312aa6460fab4cab0c5 WatchSource:0}: Error finding container f46a58ed35edf7f8e88a8cbc53e692f6b9c389d8006a9312aa6460fab4cab0c5: Status 404 returned error can't find the container with id f46a58ed35edf7f8e88a8cbc53e692f6b9c389d8006a9312aa6460fab4cab0c5 Dec 10 09:44:15 crc kubenswrapper[4715]: I1210 09:44:15.062119 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-mmntd" event={"ID":"7a38a68d-c3fe-40f5-be5d-4c75b919c4c4","Type":"ContainerStarted","Data":"4ee3d6132aaf3fde773af12cf579e368223c3bee5d4c62ea6a21d38bc4b3505a"} Dec 10 09:44:15 crc kubenswrapper[4715]: I1210 09:44:15.064139 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-cxn4s" event={"ID":"189f8af2-11e6-4140-8104-fb4a66bce776","Type":"ContainerStarted","Data":"fa63e3f400719c510b1db572627fb6be36dc281030de76e62b703fee99d220eb"} Dec 10 09:44:15 crc kubenswrapper[4715]: I1210 09:44:15.065847 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-vckw6" event={"ID":"7f4c1366-2e47-4899-a38f-55cbc26f0569","Type":"ContainerStarted","Data":"f46a58ed35edf7f8e88a8cbc53e692f6b9c389d8006a9312aa6460fab4cab0c5"} Dec 10 09:44:17 crc kubenswrapper[4715]: I1210 09:44:17.713962 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:44:17 crc kubenswrapper[4715]: I1210 09:44:17.714498 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:44:17 crc kubenswrapper[4715]: I1210 09:44:17.714571 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:44:17 crc kubenswrapper[4715]: I1210 09:44:17.715381 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed1ab5cf91267ff02613fe03163e2a0e1e6fe3ab62fe2bfb0b6ed4dcd711b4fe"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 09:44:17 crc kubenswrapper[4715]: I1210 09:44:17.715451 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://ed1ab5cf91267ff02613fe03163e2a0e1e6fe3ab62fe2bfb0b6ed4dcd711b4fe" gracePeriod=600 Dec 10 09:44:19 crc kubenswrapper[4715]: I1210 09:44:19.088874 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="ed1ab5cf91267ff02613fe03163e2a0e1e6fe3ab62fe2bfb0b6ed4dcd711b4fe" exitCode=0 Dec 10 09:44:19 crc kubenswrapper[4715]: I1210 09:44:19.088936 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"ed1ab5cf91267ff02613fe03163e2a0e1e6fe3ab62fe2bfb0b6ed4dcd711b4fe"} Dec 10 09:44:19 crc kubenswrapper[4715]: I1210 09:44:19.088975 4715 scope.go:117] "RemoveContainer" containerID="f0e5a5883c2401a24c6955974c04dfd24314a74cbb02a997f9f8a68454c4153f" Dec 10 09:44:20 crc kubenswrapper[4715]: I1210 09:44:20.095794 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-mmntd" event={"ID":"7a38a68d-c3fe-40f5-be5d-4c75b919c4c4","Type":"ContainerStarted","Data":"e0fea8c97da565564518beecddaec4ee5420d1667c5aac046813c002df64ad06"} Dec 10 09:44:20 crc kubenswrapper[4715]: I1210 09:44:20.099964 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"5d20ad65a2bc63bec106125489b213c4e46e5f275365a8160d4b290979ccc044"} Dec 10 09:44:20 crc kubenswrapper[4715]: I1210 09:44:20.118718 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-mmntd" podStartSLOduration=2.226796872 podStartE2EDuration="7.1186902s" podCreationTimestamp="2025-12-10 09:44:13 +0000 UTC" firstStartedPulling="2025-12-10 09:44:14.389654564 +0000 UTC m=+617.133200815" lastFinishedPulling="2025-12-10 09:44:19.281547892 +0000 UTC m=+622.025094143" observedRunningTime="2025-12-10 09:44:20.108135966 +0000 UTC m=+622.851682217" watchObservedRunningTime="2025-12-10 09:44:20.1186902 +0000 UTC m=+622.862236461" Dec 10 09:44:22 crc kubenswrapper[4715]: I1210 09:44:22.113437 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-vckw6" event={"ID":"7f4c1366-2e47-4899-a38f-55cbc26f0569","Type":"ContainerStarted","Data":"c5ca181fbc9998c1e997ef44e9a98689baab864d5ec69919383760b15741e876"} Dec 10 09:44:22 crc kubenswrapper[4715]: I1210 09:44:22.114244 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-vckw6" Dec 10 09:44:22 crc kubenswrapper[4715]: I1210 09:44:22.116840 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-cxn4s" event={"ID":"189f8af2-11e6-4140-8104-fb4a66bce776","Type":"ContainerStarted","Data":"a5738fade59808c553db36f3554f72ad75eaf5979350fc1f0acdbd9c9d87bdbc"} Dec 10 09:44:22 crc kubenswrapper[4715]: I1210 09:44:22.133896 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-vckw6" podStartSLOduration=2.153170054 podStartE2EDuration="9.133879659s" podCreationTimestamp="2025-12-10 09:44:13 +0000 UTC" firstStartedPulling="2025-12-10 09:44:14.650487286 +0000 UTC m=+617.394033537" lastFinishedPulling="2025-12-10 09:44:21.631196891 +0000 UTC m=+624.374743142" observedRunningTime="2025-12-10 09:44:22.133401115 +0000 UTC m=+624.876947356" watchObservedRunningTime="2025-12-10 09:44:22.133879659 +0000 UTC m=+624.877425930" Dec 10 09:44:22 crc kubenswrapper[4715]: I1210 09:44:22.155148 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-cxn4s" podStartSLOduration=2.173060105 podStartE2EDuration="9.15513268s" podCreationTimestamp="2025-12-10 09:44:13 +0000 UTC" firstStartedPulling="2025-12-10 09:44:14.649146747 +0000 UTC m=+617.392692998" lastFinishedPulling="2025-12-10 09:44:21.631219312 +0000 UTC m=+624.374765573" observedRunningTime="2025-12-10 09:44:22.153356259 +0000 UTC m=+624.896902510" watchObservedRunningTime="2025-12-10 09:44:22.15513268 +0000 UTC m=+624.898678931" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.286718 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vspkd"] Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.287432 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-acl-logging" containerID="cri-o://7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa" gracePeriod=30 Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.287485 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="northd" containerID="cri-o://6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b" gracePeriod=30 Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.287505 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="sbdb" containerID="cri-o://34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d" gracePeriod=30 Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.287540 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10" gracePeriod=30 Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.287505 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="nbdb" containerID="cri-o://0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4" gracePeriod=30 Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.287431 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-controller" containerID="cri-o://1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686" gracePeriod=30 Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.287523 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="kube-rbac-proxy-node" containerID="cri-o://a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b" gracePeriod=30 Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.336290 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" containerID="cri-o://ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f" gracePeriod=30 Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.642512 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/3.log" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.643223 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/1.log" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.645111 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.645621 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-controller/0.log" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.646062 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704441 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jcnw8"] Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704675 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-acl-logging" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704687 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-acl-logging" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704697 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704703 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704714 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="kubecfg-setup" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704722 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="kubecfg-setup" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704730 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="sbdb" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704737 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="sbdb" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704747 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="kube-rbac-proxy-node" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704753 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="kube-rbac-proxy-node" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704762 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704768 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704774 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704780 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704789 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-acl-logging" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704797 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-acl-logging" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704806 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704812 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704820 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="nbdb" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704825 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="nbdb" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704834 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704839 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704846 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704852 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.704860 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="northd" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.704867 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="northd" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705016 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705030 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-acl-logging" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705037 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705045 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="kube-rbac-proxy-node" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705053 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="northd" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705065 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="nbdb" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705073 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="sbdb" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705082 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705089 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705096 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 09:44:24 crc kubenswrapper[4715]: E1210 09:44:24.705250 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705260 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705415 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705425 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovnkube-controller" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.705435 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerName="ovn-acl-logging" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.707990 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785426 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbxh6\" (UniqueName: \"kubernetes.io/projected/2a44160c-cf30-4b13-b82c-4c402d967dd8-kube-api-access-sbxh6\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785484 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-script-lib\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785510 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785538 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-node-log\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785552 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-openvswitch\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785588 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-log-socket\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785604 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-var-lib-openvswitch\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785622 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-systemd\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785642 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-netd\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785691 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-config\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785706 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-systemd-units\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785719 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-bin\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785735 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-etc-openvswitch\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785758 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-ovn-kubernetes\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785775 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-slash\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785791 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-ovn\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785810 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovn-node-metrics-cert\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785844 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-kubelet\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785859 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-netns\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.785877 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-env-overrides\") pod \"2a44160c-cf30-4b13-b82c-4c402d967dd8\" (UID: \"2a44160c-cf30-4b13-b82c-4c402d967dd8\") " Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786111 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-slash" (OuterVolumeSpecName: "host-slash") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786161 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786168 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786185 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786186 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786205 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786210 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-node-log" (OuterVolumeSpecName: "node-log") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786225 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786250 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786257 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786275 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-log-socket" (OuterVolumeSpecName: "log-socket") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786255 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786276 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786366 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786771 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786801 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.786963 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.790769 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a44160c-cf30-4b13-b82c-4c402d967dd8-kube-api-access-sbxh6" (OuterVolumeSpecName: "kube-api-access-sbxh6") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "kube-api-access-sbxh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.790943 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.798991 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "2a44160c-cf30-4b13-b82c-4c402d967dd8" (UID: "2a44160c-cf30-4b13-b82c-4c402d967dd8"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.886981 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-run-systemd\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887044 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-etc-openvswitch\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887071 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-ovnkube-script-lib\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887095 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4kfj\" (UniqueName: \"kubernetes.io/projected/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-kube-api-access-k4kfj\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887127 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-ovn-node-metrics-cert\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887159 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-kubelet\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887187 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-slash\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887228 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-run-netns\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887281 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-var-lib-openvswitch\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887480 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-log-socket\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887602 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-run-ovn\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887695 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-env-overrides\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887790 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887841 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-cni-bin\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887967 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-node-log\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.887994 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-systemd-units\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888024 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-ovnkube-config\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888059 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-run-openvswitch\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888091 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-run-ovn-kubernetes\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888238 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-cni-netd\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888369 4715 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888390 4715 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888403 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888417 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbxh6\" (UniqueName: \"kubernetes.io/projected/2a44160c-cf30-4b13-b82c-4c402d967dd8-kube-api-access-sbxh6\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888430 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888446 4715 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888458 4715 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-node-log\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888472 4715 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888486 4715 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-log-socket\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888497 4715 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888509 4715 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888519 4715 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888530 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888541 4715 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888551 4715 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888564 4715 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888576 4715 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888589 4715 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-host-slash\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888600 4715 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a44160c-cf30-4b13-b82c-4c402d967dd8-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.888611 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a44160c-cf30-4b13-b82c-4c402d967dd8-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-cni-netd\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989503 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-run-systemd\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989520 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-etc-openvswitch\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989537 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-ovnkube-script-lib\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989553 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4kfj\" (UniqueName: \"kubernetes.io/projected/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-kube-api-access-k4kfj\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989573 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-ovn-node-metrics-cert\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989587 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-kubelet\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989586 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-run-systemd\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989603 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-cni-netd\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989642 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-slash\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989609 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-slash\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989675 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-run-netns\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989718 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-var-lib-openvswitch\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989724 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-etc-openvswitch\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989719 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-run-netns\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.990214 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-var-lib-openvswitch\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.990595 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-log-socket\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.990767 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-ovnkube-script-lib\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.990767 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-run-ovn\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.990834 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-env-overrides\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.990880 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.990936 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-cni-bin\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.990974 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-systemd-units\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991012 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-node-log\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991043 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-ovnkube-config\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991077 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-run-openvswitch\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991113 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-run-ovn-kubernetes\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991112 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-cni-bin\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991267 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-systemd-units\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991310 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-node-log\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991547 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-env-overrides\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991610 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-run-ovn-kubernetes\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991558 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-run-ovn\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991752 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991726 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-log-socket\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991809 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-run-openvswitch\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.989798 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-host-kubelet\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.991943 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-ovnkube-config\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:24 crc kubenswrapper[4715]: I1210 09:44:24.996470 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-ovn-node-metrics-cert\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.008020 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4kfj\" (UniqueName: \"kubernetes.io/projected/9e9f9542-1d87-49bb-a79f-82f7aaa50b5d-kube-api-access-k4kfj\") pod \"ovnkube-node-jcnw8\" (UID: \"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d\") " pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.020512 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:25 crc kubenswrapper[4715]: W1210 09:44:25.040242 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e9f9542_1d87_49bb_a79f_82f7aaa50b5d.slice/crio-e4d49b935b21bdd9ad9351a1c38bdd3969dd410f9c5e2a3bf9a4cae071b5bbf9 WatchSource:0}: Error finding container e4d49b935b21bdd9ad9351a1c38bdd3969dd410f9c5e2a3bf9a4cae071b5bbf9: Status 404 returned error can't find the container with id e4d49b935b21bdd9ad9351a1c38bdd3969dd410f9c5e2a3bf9a4cae071b5bbf9 Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.136539 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" event={"ID":"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d","Type":"ContainerStarted","Data":"e4d49b935b21bdd9ad9351a1c38bdd3969dd410f9c5e2a3bf9a4cae071b5bbf9"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.141003 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovnkube-controller/3.log" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.141488 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/1.log" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.144889 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-acl-logging/0.log" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.145424 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vspkd_2a44160c-cf30-4b13-b82c-4c402d967dd8/ovn-controller/0.log" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.145791 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f" exitCode=0 Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.145928 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa" exitCode=143 Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146007 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d" exitCode=0 Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146072 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4" exitCode=0 Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146132 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b" exitCode=0 Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146354 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10" exitCode=0 Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146424 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b" exitCode=0 Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146481 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a44160c-cf30-4b13-b82c-4c402d967dd8" containerID="1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686" exitCode=143 Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146273 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146670 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146750 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146882 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146978 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146854 4715 scope.go:117] "RemoveContainer" containerID="ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147061 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147219 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147243 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147250 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147270 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.146337 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147298 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147399 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147407 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147413 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147419 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147425 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147432 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147439 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147447 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147454 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147461 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147473 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147486 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147496 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147502 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147509 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147517 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147524 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147530 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147537 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147544 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147551 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147557 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147569 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147583 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147591 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147598 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147604 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147610 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147659 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147667 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147673 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147679 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147684 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147690 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147699 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vspkd" event={"ID":"2a44160c-cf30-4b13-b82c-4c402d967dd8","Type":"ContainerDied","Data":"9761e111d4530a07bf0cdc6c800b0cd4dd70e66cd1972fbd1ef86961b3879cf0"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147710 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147743 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147750 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147757 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147764 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147790 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147851 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147946 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.147959 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.148003 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.148012 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.153385 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/2.log" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.154006 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/1.log" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.154050 4715 generic.go:334] "Generic (PLEG): container finished" podID="8870da50-36bc-413e-96ba-89740d93d5d9" containerID="9262d98d872c33b9676c35f7142bb4e8274eee8158ccc7e65af9ba5eeeb62c6f" exitCode=2 Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.154085 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-98zp6" event={"ID":"8870da50-36bc-413e-96ba-89740d93d5d9","Type":"ContainerDied","Data":"9262d98d872c33b9676c35f7142bb4e8274eee8158ccc7e65af9ba5eeeb62c6f"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.154112 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b"} Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.154583 4715 scope.go:117] "RemoveContainer" containerID="9262d98d872c33b9676c35f7142bb4e8274eee8158ccc7e65af9ba5eeeb62c6f" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.155715 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-98zp6_openshift-multus(8870da50-36bc-413e-96ba-89740d93d5d9)\"" pod="openshift-multus/multus-98zp6" podUID="8870da50-36bc-413e-96ba-89740d93d5d9" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.199176 4715 scope.go:117] "RemoveContainer" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.214239 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vspkd"] Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.218254 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vspkd"] Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.224283 4715 scope.go:117] "RemoveContainer" containerID="7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.236433 4715 scope.go:117] "RemoveContainer" containerID="34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.283389 4715 scope.go:117] "RemoveContainer" containerID="0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.298063 4715 scope.go:117] "RemoveContainer" containerID="6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.311030 4715 scope.go:117] "RemoveContainer" containerID="a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.324516 4715 scope.go:117] "RemoveContainer" containerID="a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.340410 4715 scope.go:117] "RemoveContainer" containerID="e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.355239 4715 scope.go:117] "RemoveContainer" containerID="1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.370075 4715 scope.go:117] "RemoveContainer" containerID="41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.386532 4715 scope.go:117] "RemoveContainer" containerID="ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.387187 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f\": container with ID starting with ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f not found: ID does not exist" containerID="ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.387264 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f"} err="failed to get container status \"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f\": rpc error: code = NotFound desc = could not find container \"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f\": container with ID starting with ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.387309 4715 scope.go:117] "RemoveContainer" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.387742 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\": container with ID starting with 61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103 not found: ID does not exist" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.387774 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103"} err="failed to get container status \"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\": rpc error: code = NotFound desc = could not find container \"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\": container with ID starting with 61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.387802 4715 scope.go:117] "RemoveContainer" containerID="7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.388218 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\": container with ID starting with 7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa not found: ID does not exist" containerID="7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.388240 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa"} err="failed to get container status \"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\": rpc error: code = NotFound desc = could not find container \"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\": container with ID starting with 7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.388255 4715 scope.go:117] "RemoveContainer" containerID="34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.389165 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\": container with ID starting with 34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d not found: ID does not exist" containerID="34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.389303 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d"} err="failed to get container status \"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\": rpc error: code = NotFound desc = could not find container \"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\": container with ID starting with 34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.389408 4715 scope.go:117] "RemoveContainer" containerID="0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.389857 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\": container with ID starting with 0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4 not found: ID does not exist" containerID="0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.389882 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4"} err="failed to get container status \"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\": rpc error: code = NotFound desc = could not find container \"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\": container with ID starting with 0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.389945 4715 scope.go:117] "RemoveContainer" containerID="6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.390343 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\": container with ID starting with 6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b not found: ID does not exist" containerID="6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.390376 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b"} err="failed to get container status \"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\": rpc error: code = NotFound desc = could not find container \"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\": container with ID starting with 6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.390396 4715 scope.go:117] "RemoveContainer" containerID="a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.390686 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\": container with ID starting with a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10 not found: ID does not exist" containerID="a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.390714 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10"} err="failed to get container status \"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\": rpc error: code = NotFound desc = could not find container \"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\": container with ID starting with a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.390729 4715 scope.go:117] "RemoveContainer" containerID="a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.391031 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\": container with ID starting with a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b not found: ID does not exist" containerID="a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.391051 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b"} err="failed to get container status \"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\": rpc error: code = NotFound desc = could not find container \"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\": container with ID starting with a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.391064 4715 scope.go:117] "RemoveContainer" containerID="e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.391475 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\": container with ID starting with e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178 not found: ID does not exist" containerID="e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.391498 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178"} err="failed to get container status \"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\": rpc error: code = NotFound desc = could not find container \"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\": container with ID starting with e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.391516 4715 scope.go:117] "RemoveContainer" containerID="1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.391781 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\": container with ID starting with 1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686 not found: ID does not exist" containerID="1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.391878 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686"} err="failed to get container status \"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\": rpc error: code = NotFound desc = could not find container \"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\": container with ID starting with 1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.391983 4715 scope.go:117] "RemoveContainer" containerID="41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06" Dec 10 09:44:25 crc kubenswrapper[4715]: E1210 09:44:25.392726 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\": container with ID starting with 41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06 not found: ID does not exist" containerID="41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.392767 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06"} err="failed to get container status \"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\": rpc error: code = NotFound desc = could not find container \"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\": container with ID starting with 41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.392789 4715 scope.go:117] "RemoveContainer" containerID="ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.393227 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f"} err="failed to get container status \"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f\": rpc error: code = NotFound desc = could not find container \"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f\": container with ID starting with ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.393312 4715 scope.go:117] "RemoveContainer" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.395193 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103"} err="failed to get container status \"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\": rpc error: code = NotFound desc = could not find container \"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\": container with ID starting with 61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.395225 4715 scope.go:117] "RemoveContainer" containerID="7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.395461 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa"} err="failed to get container status \"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\": rpc error: code = NotFound desc = could not find container \"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\": container with ID starting with 7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.395485 4715 scope.go:117] "RemoveContainer" containerID="34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.396400 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d"} err="failed to get container status \"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\": rpc error: code = NotFound desc = could not find container \"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\": container with ID starting with 34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.396582 4715 scope.go:117] "RemoveContainer" containerID="0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.397049 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4"} err="failed to get container status \"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\": rpc error: code = NotFound desc = could not find container \"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\": container with ID starting with 0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.397215 4715 scope.go:117] "RemoveContainer" containerID="6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.397649 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b"} err="failed to get container status \"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\": rpc error: code = NotFound desc = could not find container \"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\": container with ID starting with 6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.397806 4715 scope.go:117] "RemoveContainer" containerID="a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.398348 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10"} err="failed to get container status \"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\": rpc error: code = NotFound desc = could not find container \"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\": container with ID starting with a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.398455 4715 scope.go:117] "RemoveContainer" containerID="a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.399458 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b"} err="failed to get container status \"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\": rpc error: code = NotFound desc = could not find container \"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\": container with ID starting with a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.399679 4715 scope.go:117] "RemoveContainer" containerID="e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.400583 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178"} err="failed to get container status \"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\": rpc error: code = NotFound desc = could not find container \"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\": container with ID starting with e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.400617 4715 scope.go:117] "RemoveContainer" containerID="1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.401451 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686"} err="failed to get container status \"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\": rpc error: code = NotFound desc = could not find container \"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\": container with ID starting with 1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.401693 4715 scope.go:117] "RemoveContainer" containerID="41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.405220 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06"} err="failed to get container status \"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\": rpc error: code = NotFound desc = could not find container \"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\": container with ID starting with 41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.405296 4715 scope.go:117] "RemoveContainer" containerID="ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.405683 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f"} err="failed to get container status \"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f\": rpc error: code = NotFound desc = could not find container \"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f\": container with ID starting with ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.405814 4715 scope.go:117] "RemoveContainer" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.406823 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103"} err="failed to get container status \"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\": rpc error: code = NotFound desc = could not find container \"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\": container with ID starting with 61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.406967 4715 scope.go:117] "RemoveContainer" containerID="7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.408065 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa"} err="failed to get container status \"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\": rpc error: code = NotFound desc = could not find container \"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\": container with ID starting with 7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.408130 4715 scope.go:117] "RemoveContainer" containerID="34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.408516 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d"} err="failed to get container status \"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\": rpc error: code = NotFound desc = could not find container \"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\": container with ID starting with 34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.408647 4715 scope.go:117] "RemoveContainer" containerID="0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.409115 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4"} err="failed to get container status \"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\": rpc error: code = NotFound desc = could not find container \"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\": container with ID starting with 0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.409151 4715 scope.go:117] "RemoveContainer" containerID="6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.409514 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b"} err="failed to get container status \"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\": rpc error: code = NotFound desc = could not find container \"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\": container with ID starting with 6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.409541 4715 scope.go:117] "RemoveContainer" containerID="a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.409789 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10"} err="failed to get container status \"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\": rpc error: code = NotFound desc = could not find container \"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\": container with ID starting with a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.409810 4715 scope.go:117] "RemoveContainer" containerID="a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.410296 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b"} err="failed to get container status \"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\": rpc error: code = NotFound desc = could not find container \"a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b\": container with ID starting with a6f13eeb7b0e40a7cca5a38ea9ba9d5c53406a0b4bc67ea937b4f52b2abfbd5b not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.410328 4715 scope.go:117] "RemoveContainer" containerID="e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.410804 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178"} err="failed to get container status \"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\": rpc error: code = NotFound desc = could not find container \"e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178\": container with ID starting with e2c1d71f90414f9e5e75a11bd854569981ed22b8e9023f45b29c99cd8ed56178 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.410830 4715 scope.go:117] "RemoveContainer" containerID="1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.411053 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686"} err="failed to get container status \"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\": rpc error: code = NotFound desc = could not find container \"1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686\": container with ID starting with 1942e83dce938f422006d081c293edeabdabdba6539fca3cd58888af5c96f686 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.411082 4715 scope.go:117] "RemoveContainer" containerID="41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.411335 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06"} err="failed to get container status \"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\": rpc error: code = NotFound desc = could not find container \"41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06\": container with ID starting with 41984b9f8a506ab31f29a697a7480ec2639f7c1ff159a4020b93c1b11ecb5f06 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.411429 4715 scope.go:117] "RemoveContainer" containerID="ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.411791 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f"} err="failed to get container status \"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f\": rpc error: code = NotFound desc = could not find container \"ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f\": container with ID starting with ebff236bdde1b2e16a43f3f165011a0a9e242fff3ca8a157e1dfdf191d217f1f not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.411817 4715 scope.go:117] "RemoveContainer" containerID="61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.412078 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103"} err="failed to get container status \"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\": rpc error: code = NotFound desc = could not find container \"61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103\": container with ID starting with 61da0f5c789fae8adc30ad69300de4c3fa7c6ceb800c32a69230fa9b48eba103 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.412172 4715 scope.go:117] "RemoveContainer" containerID="7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.412494 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa"} err="failed to get container status \"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\": rpc error: code = NotFound desc = could not find container \"7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa\": container with ID starting with 7291d5c8b6ba32b057e81f1b26d1037f4fceb18a97461f13528220db1bd9bcfa not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.412518 4715 scope.go:117] "RemoveContainer" containerID="34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.412745 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d"} err="failed to get container status \"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\": rpc error: code = NotFound desc = could not find container \"34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d\": container with ID starting with 34f6b1d9f1b6ffc9eb2440e9ced1ab19c675217ede167e314cf930940a8c937d not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.412849 4715 scope.go:117] "RemoveContainer" containerID="0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.413268 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4"} err="failed to get container status \"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\": rpc error: code = NotFound desc = could not find container \"0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4\": container with ID starting with 0f2070b929dd763a5e8e68d843b9494ab4c2808e514be704df6247453a5e37c4 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.413297 4715 scope.go:117] "RemoveContainer" containerID="6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.413514 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b"} err="failed to get container status \"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\": rpc error: code = NotFound desc = could not find container \"6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b\": container with ID starting with 6c60c5cd871e3d23768db157106e31fa76b8f4f0c6d4275cf5ffa4c2fcafd80b not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.413533 4715 scope.go:117] "RemoveContainer" containerID="a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.413778 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10"} err="failed to get container status \"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\": rpc error: code = NotFound desc = could not find container \"a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10\": container with ID starting with a1ec7b21934f16e4b3b422dfccce09860673745c838bbcf34405a07a67b5dc10 not found: ID does not exist" Dec 10 09:44:25 crc kubenswrapper[4715]: I1210 09:44:25.615992 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a44160c-cf30-4b13-b82c-4c402d967dd8" path="/var/lib/kubelet/pods/2a44160c-cf30-4b13-b82c-4c402d967dd8/volumes" Dec 10 09:44:26 crc kubenswrapper[4715]: I1210 09:44:26.163648 4715 generic.go:334] "Generic (PLEG): container finished" podID="9e9f9542-1d87-49bb-a79f-82f7aaa50b5d" containerID="cd98c8ae39f4253d974f7534db24b8ca777e850d1b1c389b469f9c717e1b4a11" exitCode=0 Dec 10 09:44:26 crc kubenswrapper[4715]: I1210 09:44:26.163687 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" event={"ID":"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d","Type":"ContainerDied","Data":"cd98c8ae39f4253d974f7534db24b8ca777e850d1b1c389b469f9c717e1b4a11"} Dec 10 09:44:27 crc kubenswrapper[4715]: I1210 09:44:27.173686 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" event={"ID":"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d","Type":"ContainerStarted","Data":"91a201aceba7e6536f93554c6a6266b50976dc15ba3bb82f14ac4323713949ce"} Dec 10 09:44:27 crc kubenswrapper[4715]: I1210 09:44:27.174295 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" event={"ID":"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d","Type":"ContainerStarted","Data":"8ff1f4f157ae154ab9913ee48f98f09ea85068f28541eb8a9a15c438dc2075bb"} Dec 10 09:44:27 crc kubenswrapper[4715]: I1210 09:44:27.174316 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" event={"ID":"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d","Type":"ContainerStarted","Data":"1b843b59ba2fa48e19547f7cc6495511362f9057228eb5da619b7f14fd686c7d"} Dec 10 09:44:27 crc kubenswrapper[4715]: I1210 09:44:27.174330 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" event={"ID":"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d","Type":"ContainerStarted","Data":"3de37a3e6019a7d349cd0334ef7b0a878aa97ea91c49b55ae5be1d53480e9bf4"} Dec 10 09:44:27 crc kubenswrapper[4715]: I1210 09:44:27.174341 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" event={"ID":"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d","Type":"ContainerStarted","Data":"cfb647811ef5434bdf4b617bf638ded2febfc6db2194cde4bae87506c68ca7ea"} Dec 10 09:44:27 crc kubenswrapper[4715]: I1210 09:44:27.174353 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" event={"ID":"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d","Type":"ContainerStarted","Data":"f37b4b2443f170a515c6e71c2d17e99f8c48b55d0d6a60cb63ba9432b78f4841"} Dec 10 09:44:29 crc kubenswrapper[4715]: I1210 09:44:29.148430 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-vckw6" Dec 10 09:44:29 crc kubenswrapper[4715]: I1210 09:44:29.186799 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" event={"ID":"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d","Type":"ContainerStarted","Data":"eae0f7ef24bfd7d7db80d5dc32cbc80a00e06267abedb51f3c936a7833243c63"} Dec 10 09:44:32 crc kubenswrapper[4715]: I1210 09:44:32.210229 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" event={"ID":"9e9f9542-1d87-49bb-a79f-82f7aaa50b5d","Type":"ContainerStarted","Data":"7073f6af066e8f35aecc6a903074a71d681f18d303f7e9b9cb8a6eb11e484953"} Dec 10 09:44:32 crc kubenswrapper[4715]: I1210 09:44:32.210739 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:32 crc kubenswrapper[4715]: I1210 09:44:32.211064 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:32 crc kubenswrapper[4715]: I1210 09:44:32.246561 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" podStartSLOduration=8.246543235 podStartE2EDuration="8.246543235s" podCreationTimestamp="2025-12-10 09:44:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:44:32.241742216 +0000 UTC m=+634.985288467" watchObservedRunningTime="2025-12-10 09:44:32.246543235 +0000 UTC m=+634.990089486" Dec 10 09:44:32 crc kubenswrapper[4715]: I1210 09:44:32.250045 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:33 crc kubenswrapper[4715]: I1210 09:44:33.216284 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:33 crc kubenswrapper[4715]: I1210 09:44:33.279210 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:38 crc kubenswrapper[4715]: I1210 09:44:38.604804 4715 scope.go:117] "RemoveContainer" containerID="9262d98d872c33b9676c35f7142bb4e8274eee8158ccc7e65af9ba5eeeb62c6f" Dec 10 09:44:38 crc kubenswrapper[4715]: E1210 09:44:38.605371 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-98zp6_openshift-multus(8870da50-36bc-413e-96ba-89740d93d5d9)\"" pod="openshift-multus/multus-98zp6" podUID="8870da50-36bc-413e-96ba-89740d93d5d9" Dec 10 09:44:53 crc kubenswrapper[4715]: I1210 09:44:53.605377 4715 scope.go:117] "RemoveContainer" containerID="9262d98d872c33b9676c35f7142bb4e8274eee8158ccc7e65af9ba5eeeb62c6f" Dec 10 09:44:54 crc kubenswrapper[4715]: I1210 09:44:54.339713 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/2.log" Dec 10 09:44:54 crc kubenswrapper[4715]: I1210 09:44:54.341065 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/1.log" Dec 10 09:44:54 crc kubenswrapper[4715]: I1210 09:44:54.341152 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-98zp6" event={"ID":"8870da50-36bc-413e-96ba-89740d93d5d9","Type":"ContainerStarted","Data":"b28612c9bf6ad763652e01b8e3ab39a9a51bc08f947771fa2eb395c699fa89dc"} Dec 10 09:44:55 crc kubenswrapper[4715]: I1210 09:44:55.146559 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jcnw8" Dec 10 09:44:57 crc kubenswrapper[4715]: I1210 09:44:57.866838 4715 scope.go:117] "RemoveContainer" containerID="3347ecc1a08611a8d65de9cbdec1f6ebc5dde059577cbd70c1a34d121d044b7b" Dec 10 09:44:58 crc kubenswrapper[4715]: I1210 09:44:58.368110 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-98zp6_8870da50-36bc-413e-96ba-89740d93d5d9/kube-multus/2.log" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.180744 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq"] Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.181636 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.183967 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.184009 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.188589 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq"] Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.194716 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e404a964-ca8d-49f5-a026-eeea3c71582c-secret-volume\") pod \"collect-profiles-29422665-j4qmq\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.194777 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e404a964-ca8d-49f5-a026-eeea3c71582c-config-volume\") pod \"collect-profiles-29422665-j4qmq\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.194849 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnnnq\" (UniqueName: \"kubernetes.io/projected/e404a964-ca8d-49f5-a026-eeea3c71582c-kube-api-access-xnnnq\") pod \"collect-profiles-29422665-j4qmq\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.295468 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e404a964-ca8d-49f5-a026-eeea3c71582c-config-volume\") pod \"collect-profiles-29422665-j4qmq\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.295587 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnnnq\" (UniqueName: \"kubernetes.io/projected/e404a964-ca8d-49f5-a026-eeea3c71582c-kube-api-access-xnnnq\") pod \"collect-profiles-29422665-j4qmq\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.295616 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e404a964-ca8d-49f5-a026-eeea3c71582c-secret-volume\") pod \"collect-profiles-29422665-j4qmq\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.296433 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e404a964-ca8d-49f5-a026-eeea3c71582c-config-volume\") pod \"collect-profiles-29422665-j4qmq\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.300847 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e404a964-ca8d-49f5-a026-eeea3c71582c-secret-volume\") pod \"collect-profiles-29422665-j4qmq\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.312190 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnnnq\" (UniqueName: \"kubernetes.io/projected/e404a964-ca8d-49f5-a026-eeea3c71582c-kube-api-access-xnnnq\") pod \"collect-profiles-29422665-j4qmq\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.499246 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:00 crc kubenswrapper[4715]: I1210 09:45:00.694484 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq"] Dec 10 09:45:00 crc kubenswrapper[4715]: W1210 09:45:00.698267 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode404a964_ca8d_49f5_a026_eeea3c71582c.slice/crio-bd9552843ab253538cc7d1e2374ae58894cf80903d604dee1ba5559f36af8772 WatchSource:0}: Error finding container bd9552843ab253538cc7d1e2374ae58894cf80903d604dee1ba5559f36af8772: Status 404 returned error can't find the container with id bd9552843ab253538cc7d1e2374ae58894cf80903d604dee1ba5559f36af8772 Dec 10 09:45:01 crc kubenswrapper[4715]: I1210 09:45:01.384060 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" event={"ID":"e404a964-ca8d-49f5-a026-eeea3c71582c","Type":"ContainerStarted","Data":"bd9552843ab253538cc7d1e2374ae58894cf80903d604dee1ba5559f36af8772"} Dec 10 09:45:02 crc kubenswrapper[4715]: I1210 09:45:02.391097 4715 generic.go:334] "Generic (PLEG): container finished" podID="e404a964-ca8d-49f5-a026-eeea3c71582c" containerID="b87daa8256509516b33fbe9f34846c8af055f3366fd79998bbd45f5a0fa2a5ed" exitCode=0 Dec 10 09:45:02 crc kubenswrapper[4715]: I1210 09:45:02.391139 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" event={"ID":"e404a964-ca8d-49f5-a026-eeea3c71582c","Type":"ContainerDied","Data":"b87daa8256509516b33fbe9f34846c8af055f3366fd79998bbd45f5a0fa2a5ed"} Dec 10 09:45:03 crc kubenswrapper[4715]: I1210 09:45:03.613710 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:03 crc kubenswrapper[4715]: I1210 09:45:03.739944 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e404a964-ca8d-49f5-a026-eeea3c71582c-config-volume\") pod \"e404a964-ca8d-49f5-a026-eeea3c71582c\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " Dec 10 09:45:03 crc kubenswrapper[4715]: I1210 09:45:03.740408 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e404a964-ca8d-49f5-a026-eeea3c71582c-secret-volume\") pod \"e404a964-ca8d-49f5-a026-eeea3c71582c\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " Dec 10 09:45:03 crc kubenswrapper[4715]: I1210 09:45:03.740454 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnnnq\" (UniqueName: \"kubernetes.io/projected/e404a964-ca8d-49f5-a026-eeea3c71582c-kube-api-access-xnnnq\") pod \"e404a964-ca8d-49f5-a026-eeea3c71582c\" (UID: \"e404a964-ca8d-49f5-a026-eeea3c71582c\") " Dec 10 09:45:03 crc kubenswrapper[4715]: I1210 09:45:03.740774 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e404a964-ca8d-49f5-a026-eeea3c71582c-config-volume" (OuterVolumeSpecName: "config-volume") pod "e404a964-ca8d-49f5-a026-eeea3c71582c" (UID: "e404a964-ca8d-49f5-a026-eeea3c71582c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:45:03 crc kubenswrapper[4715]: I1210 09:45:03.745625 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e404a964-ca8d-49f5-a026-eeea3c71582c-kube-api-access-xnnnq" (OuterVolumeSpecName: "kube-api-access-xnnnq") pod "e404a964-ca8d-49f5-a026-eeea3c71582c" (UID: "e404a964-ca8d-49f5-a026-eeea3c71582c"). InnerVolumeSpecName "kube-api-access-xnnnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:45:03 crc kubenswrapper[4715]: I1210 09:45:03.745873 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e404a964-ca8d-49f5-a026-eeea3c71582c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e404a964-ca8d-49f5-a026-eeea3c71582c" (UID: "e404a964-ca8d-49f5-a026-eeea3c71582c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:45:03 crc kubenswrapper[4715]: I1210 09:45:03.841213 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e404a964-ca8d-49f5-a026-eeea3c71582c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 09:45:03 crc kubenswrapper[4715]: I1210 09:45:03.841463 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e404a964-ca8d-49f5-a026-eeea3c71582c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 09:45:03 crc kubenswrapper[4715]: I1210 09:45:03.841535 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnnnq\" (UniqueName: \"kubernetes.io/projected/e404a964-ca8d-49f5-a026-eeea3c71582c-kube-api-access-xnnnq\") on node \"crc\" DevicePath \"\"" Dec 10 09:45:04 crc kubenswrapper[4715]: I1210 09:45:04.406844 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" event={"ID":"e404a964-ca8d-49f5-a026-eeea3c71582c","Type":"ContainerDied","Data":"bd9552843ab253538cc7d1e2374ae58894cf80903d604dee1ba5559f36af8772"} Dec 10 09:45:04 crc kubenswrapper[4715]: I1210 09:45:04.406905 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd9552843ab253538cc7d1e2374ae58894cf80903d604dee1ba5559f36af8772" Dec 10 09:45:04 crc kubenswrapper[4715]: I1210 09:45:04.406865 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.007055 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t"] Dec 10 09:45:15 crc kubenswrapper[4715]: E1210 09:45:15.007841 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e404a964-ca8d-49f5-a026-eeea3c71582c" containerName="collect-profiles" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.007856 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e404a964-ca8d-49f5-a026-eeea3c71582c" containerName="collect-profiles" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.008002 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e404a964-ca8d-49f5-a026-eeea3c71582c" containerName="collect-profiles" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.008958 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.012056 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.015599 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t"] Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.181023 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.181159 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qvgs\" (UniqueName: \"kubernetes.io/projected/580ed0cb-7745-46e9-8437-b1e0ba6199cd-kube-api-access-8qvgs\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.181283 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.282706 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.282814 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.282867 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qvgs\" (UniqueName: \"kubernetes.io/projected/580ed0cb-7745-46e9-8437-b1e0ba6199cd-kube-api-access-8qvgs\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.283527 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.283962 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.314217 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qvgs\" (UniqueName: \"kubernetes.io/projected/580ed0cb-7745-46e9-8437-b1e0ba6199cd-kube-api-access-8qvgs\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.321971 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:15 crc kubenswrapper[4715]: I1210 09:45:15.541908 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t"] Dec 10 09:45:16 crc kubenswrapper[4715]: I1210 09:45:16.465834 4715 generic.go:334] "Generic (PLEG): container finished" podID="580ed0cb-7745-46e9-8437-b1e0ba6199cd" containerID="1664d425a701a496c4fa95fac6cf2ad0129ecf30ca8933bf7139198c0bc83304" exitCode=0 Dec 10 09:45:16 crc kubenswrapper[4715]: I1210 09:45:16.465873 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" event={"ID":"580ed0cb-7745-46e9-8437-b1e0ba6199cd","Type":"ContainerDied","Data":"1664d425a701a496c4fa95fac6cf2ad0129ecf30ca8933bf7139198c0bc83304"} Dec 10 09:45:16 crc kubenswrapper[4715]: I1210 09:45:16.465893 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" event={"ID":"580ed0cb-7745-46e9-8437-b1e0ba6199cd","Type":"ContainerStarted","Data":"d5792bebaff5c92f1335d8984bc874786c3764d4394cc1093a96b91886be458e"} Dec 10 09:45:18 crc kubenswrapper[4715]: I1210 09:45:18.480558 4715 generic.go:334] "Generic (PLEG): container finished" podID="580ed0cb-7745-46e9-8437-b1e0ba6199cd" containerID="070c6e49355c5695dc600240b21a69a6cd874d0bcc3fc8442f7b9648d9a4c079" exitCode=0 Dec 10 09:45:18 crc kubenswrapper[4715]: I1210 09:45:18.480635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" event={"ID":"580ed0cb-7745-46e9-8437-b1e0ba6199cd","Type":"ContainerDied","Data":"070c6e49355c5695dc600240b21a69a6cd874d0bcc3fc8442f7b9648d9a4c079"} Dec 10 09:45:19 crc kubenswrapper[4715]: I1210 09:45:19.489800 4715 generic.go:334] "Generic (PLEG): container finished" podID="580ed0cb-7745-46e9-8437-b1e0ba6199cd" containerID="0b567ea1a15a7f90d44f0120fb0b050d42a1446e8664b2d26e3b8d99443dc7d0" exitCode=0 Dec 10 09:45:19 crc kubenswrapper[4715]: I1210 09:45:19.489930 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" event={"ID":"580ed0cb-7745-46e9-8437-b1e0ba6199cd","Type":"ContainerDied","Data":"0b567ea1a15a7f90d44f0120fb0b050d42a1446e8664b2d26e3b8d99443dc7d0"} Dec 10 09:45:20 crc kubenswrapper[4715]: I1210 09:45:20.734153 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:20 crc kubenswrapper[4715]: I1210 09:45:20.851323 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-util\") pod \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " Dec 10 09:45:20 crc kubenswrapper[4715]: I1210 09:45:20.851407 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qvgs\" (UniqueName: \"kubernetes.io/projected/580ed0cb-7745-46e9-8437-b1e0ba6199cd-kube-api-access-8qvgs\") pod \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " Dec 10 09:45:20 crc kubenswrapper[4715]: I1210 09:45:20.851493 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-bundle\") pod \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\" (UID: \"580ed0cb-7745-46e9-8437-b1e0ba6199cd\") " Dec 10 09:45:20 crc kubenswrapper[4715]: I1210 09:45:20.852084 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-bundle" (OuterVolumeSpecName: "bundle") pod "580ed0cb-7745-46e9-8437-b1e0ba6199cd" (UID: "580ed0cb-7745-46e9-8437-b1e0ba6199cd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:45:20 crc kubenswrapper[4715]: I1210 09:45:20.852394 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:45:20 crc kubenswrapper[4715]: I1210 09:45:20.857677 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/580ed0cb-7745-46e9-8437-b1e0ba6199cd-kube-api-access-8qvgs" (OuterVolumeSpecName: "kube-api-access-8qvgs") pod "580ed0cb-7745-46e9-8437-b1e0ba6199cd" (UID: "580ed0cb-7745-46e9-8437-b1e0ba6199cd"). InnerVolumeSpecName "kube-api-access-8qvgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:45:20 crc kubenswrapper[4715]: I1210 09:45:20.869679 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-util" (OuterVolumeSpecName: "util") pod "580ed0cb-7745-46e9-8437-b1e0ba6199cd" (UID: "580ed0cb-7745-46e9-8437-b1e0ba6199cd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:45:20 crc kubenswrapper[4715]: I1210 09:45:20.953677 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qvgs\" (UniqueName: \"kubernetes.io/projected/580ed0cb-7745-46e9-8437-b1e0ba6199cd-kube-api-access-8qvgs\") on node \"crc\" DevicePath \"\"" Dec 10 09:45:20 crc kubenswrapper[4715]: I1210 09:45:20.953717 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/580ed0cb-7745-46e9-8437-b1e0ba6199cd-util\") on node \"crc\" DevicePath \"\"" Dec 10 09:45:21 crc kubenswrapper[4715]: I1210 09:45:21.507443 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" event={"ID":"580ed0cb-7745-46e9-8437-b1e0ba6199cd","Type":"ContainerDied","Data":"d5792bebaff5c92f1335d8984bc874786c3764d4394cc1093a96b91886be458e"} Dec 10 09:45:21 crc kubenswrapper[4715]: I1210 09:45:21.507515 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5792bebaff5c92f1335d8984bc874786c3764d4394cc1093a96b91886be458e" Dec 10 09:45:21 crc kubenswrapper[4715]: I1210 09:45:21.507863 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t" Dec 10 09:45:22 crc kubenswrapper[4715]: I1210 09:45:22.957104 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww"] Dec 10 09:45:22 crc kubenswrapper[4715]: E1210 09:45:22.957294 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="580ed0cb-7745-46e9-8437-b1e0ba6199cd" containerName="util" Dec 10 09:45:22 crc kubenswrapper[4715]: I1210 09:45:22.957304 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="580ed0cb-7745-46e9-8437-b1e0ba6199cd" containerName="util" Dec 10 09:45:22 crc kubenswrapper[4715]: E1210 09:45:22.957315 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="580ed0cb-7745-46e9-8437-b1e0ba6199cd" containerName="pull" Dec 10 09:45:22 crc kubenswrapper[4715]: I1210 09:45:22.957321 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="580ed0cb-7745-46e9-8437-b1e0ba6199cd" containerName="pull" Dec 10 09:45:22 crc kubenswrapper[4715]: E1210 09:45:22.957334 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="580ed0cb-7745-46e9-8437-b1e0ba6199cd" containerName="extract" Dec 10 09:45:22 crc kubenswrapper[4715]: I1210 09:45:22.957340 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="580ed0cb-7745-46e9-8437-b1e0ba6199cd" containerName="extract" Dec 10 09:45:22 crc kubenswrapper[4715]: I1210 09:45:22.957454 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="580ed0cb-7745-46e9-8437-b1e0ba6199cd" containerName="extract" Dec 10 09:45:22 crc kubenswrapper[4715]: I1210 09:45:22.957819 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww" Dec 10 09:45:22 crc kubenswrapper[4715]: I1210 09:45:22.963106 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 10 09:45:22 crc kubenswrapper[4715]: I1210 09:45:22.963160 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 10 09:45:22 crc kubenswrapper[4715]: I1210 09:45:22.963730 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-58bt9" Dec 10 09:45:22 crc kubenswrapper[4715]: I1210 09:45:22.976440 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww"] Dec 10 09:45:23 crc kubenswrapper[4715]: I1210 09:45:23.081092 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxhn2\" (UniqueName: \"kubernetes.io/projected/bb6d3ad9-978f-48ff-b72c-434fa2c5783e-kube-api-access-dxhn2\") pod \"nmstate-operator-5b5b58f5c8-bztww\" (UID: \"bb6d3ad9-978f-48ff-b72c-434fa2c5783e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww" Dec 10 09:45:23 crc kubenswrapper[4715]: I1210 09:45:23.182403 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxhn2\" (UniqueName: \"kubernetes.io/projected/bb6d3ad9-978f-48ff-b72c-434fa2c5783e-kube-api-access-dxhn2\") pod \"nmstate-operator-5b5b58f5c8-bztww\" (UID: \"bb6d3ad9-978f-48ff-b72c-434fa2c5783e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww" Dec 10 09:45:23 crc kubenswrapper[4715]: I1210 09:45:23.206929 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxhn2\" (UniqueName: \"kubernetes.io/projected/bb6d3ad9-978f-48ff-b72c-434fa2c5783e-kube-api-access-dxhn2\") pod \"nmstate-operator-5b5b58f5c8-bztww\" (UID: \"bb6d3ad9-978f-48ff-b72c-434fa2c5783e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww" Dec 10 09:45:23 crc kubenswrapper[4715]: I1210 09:45:23.272128 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww" Dec 10 09:45:23 crc kubenswrapper[4715]: I1210 09:45:23.728039 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww"] Dec 10 09:45:24 crc kubenswrapper[4715]: I1210 09:45:24.528191 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww" event={"ID":"bb6d3ad9-978f-48ff-b72c-434fa2c5783e","Type":"ContainerStarted","Data":"57625ccc5ae21e87588c4740a00797daa264e51414f40d35d481cf5db78694b5"} Dec 10 09:45:28 crc kubenswrapper[4715]: I1210 09:45:28.552522 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww" event={"ID":"bb6d3ad9-978f-48ff-b72c-434fa2c5783e","Type":"ContainerStarted","Data":"507ef698095cc60870a7ef810d734ba803903430eb86f5d0d4c5b2206c202608"} Dec 10 09:45:28 crc kubenswrapper[4715]: I1210 09:45:28.567843 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bztww" podStartSLOduration=2.451899676 podStartE2EDuration="6.567822665s" podCreationTimestamp="2025-12-10 09:45:22 +0000 UTC" firstStartedPulling="2025-12-10 09:45:23.743348553 +0000 UTC m=+686.486894804" lastFinishedPulling="2025-12-10 09:45:27.859271542 +0000 UTC m=+690.602817793" observedRunningTime="2025-12-10 09:45:28.567346282 +0000 UTC m=+691.310892543" watchObservedRunningTime="2025-12-10 09:45:28.567822665 +0000 UTC m=+691.311368916" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.524279 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z"] Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.525366 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.526727 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-bf5fm" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.543164 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47"] Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.544112 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.547957 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.560303 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-hhxxc"] Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.561162 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.566831 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47"] Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.590691 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bc0de16a-d1b8-41b0-ad21-6015ea156e9a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-nsj47\" (UID: \"bc0de16a-d1b8-41b0-ad21-6015ea156e9a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.590754 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd8kc\" (UniqueName: \"kubernetes.io/projected/bc0de16a-d1b8-41b0-ad21-6015ea156e9a-kube-api-access-hd8kc\") pod \"nmstate-webhook-5f6d4c5ccb-nsj47\" (UID: \"bc0de16a-d1b8-41b0-ad21-6015ea156e9a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.590784 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/11a0c637-5b67-4db5-8fd0-33c6fabc6103-ovs-socket\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.590847 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/11a0c637-5b67-4db5-8fd0-33c6fabc6103-nmstate-lock\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.590865 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzk6z\" (UniqueName: \"kubernetes.io/projected/adb59670-27dd-447f-8ccc-ca87956810ec-kube-api-access-rzk6z\") pod \"nmstate-metrics-7f946cbc9-lwx7z\" (UID: \"adb59670-27dd-447f-8ccc-ca87956810ec\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.590897 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz6b7\" (UniqueName: \"kubernetes.io/projected/11a0c637-5b67-4db5-8fd0-33c6fabc6103-kube-api-access-vz6b7\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.590932 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/11a0c637-5b67-4db5-8fd0-33c6fabc6103-dbus-socket\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.591516 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z"] Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.660639 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc"] Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.661523 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.663748 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9nvrx" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.663773 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.663897 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.674154 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc"] Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.691656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzk6z\" (UniqueName: \"kubernetes.io/projected/adb59670-27dd-447f-8ccc-ca87956810ec-kube-api-access-rzk6z\") pod \"nmstate-metrics-7f946cbc9-lwx7z\" (UID: \"adb59670-27dd-447f-8ccc-ca87956810ec\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.691727 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtgvp\" (UniqueName: \"kubernetes.io/projected/2993b76a-4f33-4a02-8845-e694f2f65179-kube-api-access-wtgvp\") pod \"nmstate-console-plugin-7fbb5f6569-9srkc\" (UID: \"2993b76a-4f33-4a02-8845-e694f2f65179\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.691778 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz6b7\" (UniqueName: \"kubernetes.io/projected/11a0c637-5b67-4db5-8fd0-33c6fabc6103-kube-api-access-vz6b7\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.691809 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/11a0c637-5b67-4db5-8fd0-33c6fabc6103-dbus-socket\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.691842 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bc0de16a-d1b8-41b0-ad21-6015ea156e9a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-nsj47\" (UID: \"bc0de16a-d1b8-41b0-ad21-6015ea156e9a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.691876 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd8kc\" (UniqueName: \"kubernetes.io/projected/bc0de16a-d1b8-41b0-ad21-6015ea156e9a-kube-api-access-hd8kc\") pod \"nmstate-webhook-5f6d4c5ccb-nsj47\" (UID: \"bc0de16a-d1b8-41b0-ad21-6015ea156e9a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.691908 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2993b76a-4f33-4a02-8845-e694f2f65179-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-9srkc\" (UID: \"2993b76a-4f33-4a02-8845-e694f2f65179\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.691945 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/11a0c637-5b67-4db5-8fd0-33c6fabc6103-ovs-socket\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.691980 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2993b76a-4f33-4a02-8845-e694f2f65179-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-9srkc\" (UID: \"2993b76a-4f33-4a02-8845-e694f2f65179\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.692010 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/11a0c637-5b67-4db5-8fd0-33c6fabc6103-nmstate-lock\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.692090 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/11a0c637-5b67-4db5-8fd0-33c6fabc6103-nmstate-lock\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.692758 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/11a0c637-5b67-4db5-8fd0-33c6fabc6103-dbus-socket\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: E1210 09:45:29.692773 4715 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 10 09:45:29 crc kubenswrapper[4715]: E1210 09:45:29.692815 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc0de16a-d1b8-41b0-ad21-6015ea156e9a-tls-key-pair podName:bc0de16a-d1b8-41b0-ad21-6015ea156e9a nodeName:}" failed. No retries permitted until 2025-12-10 09:45:30.192801075 +0000 UTC m=+692.936347326 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/bc0de16a-d1b8-41b0-ad21-6015ea156e9a-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-nsj47" (UID: "bc0de16a-d1b8-41b0-ad21-6015ea156e9a") : secret "openshift-nmstate-webhook" not found Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.692866 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/11a0c637-5b67-4db5-8fd0-33c6fabc6103-ovs-socket\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.710850 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzk6z\" (UniqueName: \"kubernetes.io/projected/adb59670-27dd-447f-8ccc-ca87956810ec-kube-api-access-rzk6z\") pod \"nmstate-metrics-7f946cbc9-lwx7z\" (UID: \"adb59670-27dd-447f-8ccc-ca87956810ec\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.711058 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd8kc\" (UniqueName: \"kubernetes.io/projected/bc0de16a-d1b8-41b0-ad21-6015ea156e9a-kube-api-access-hd8kc\") pod \"nmstate-webhook-5f6d4c5ccb-nsj47\" (UID: \"bc0de16a-d1b8-41b0-ad21-6015ea156e9a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.716898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz6b7\" (UniqueName: \"kubernetes.io/projected/11a0c637-5b67-4db5-8fd0-33c6fabc6103-kube-api-access-vz6b7\") pod \"nmstate-handler-hhxxc\" (UID: \"11a0c637-5b67-4db5-8fd0-33c6fabc6103\") " pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.793094 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2993b76a-4f33-4a02-8845-e694f2f65179-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-9srkc\" (UID: \"2993b76a-4f33-4a02-8845-e694f2f65179\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.793160 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2993b76a-4f33-4a02-8845-e694f2f65179-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-9srkc\" (UID: \"2993b76a-4f33-4a02-8845-e694f2f65179\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.793210 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtgvp\" (UniqueName: \"kubernetes.io/projected/2993b76a-4f33-4a02-8845-e694f2f65179-kube-api-access-wtgvp\") pod \"nmstate-console-plugin-7fbb5f6569-9srkc\" (UID: \"2993b76a-4f33-4a02-8845-e694f2f65179\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:29 crc kubenswrapper[4715]: E1210 09:45:29.794002 4715 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 10 09:45:29 crc kubenswrapper[4715]: E1210 09:45:29.794062 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2993b76a-4f33-4a02-8845-e694f2f65179-plugin-serving-cert podName:2993b76a-4f33-4a02-8845-e694f2f65179 nodeName:}" failed. No retries permitted until 2025-12-10 09:45:30.294041535 +0000 UTC m=+693.037587786 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/2993b76a-4f33-4a02-8845-e694f2f65179-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-9srkc" (UID: "2993b76a-4f33-4a02-8845-e694f2f65179") : secret "plugin-serving-cert" not found Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.794726 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2993b76a-4f33-4a02-8845-e694f2f65179-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-9srkc\" (UID: \"2993b76a-4f33-4a02-8845-e694f2f65179\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.810095 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtgvp\" (UniqueName: \"kubernetes.io/projected/2993b76a-4f33-4a02-8845-e694f2f65179-kube-api-access-wtgvp\") pod \"nmstate-console-plugin-7fbb5f6569-9srkc\" (UID: \"2993b76a-4f33-4a02-8845-e694f2f65179\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.841872 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.871062 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6b86b8bf9d-69c54"] Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.873046 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.878480 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.885986 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6b86b8bf9d-69c54"] Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.894603 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-oauth-serving-cert\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.894672 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36af9a72-1276-40cc-b462-84af24cf5abd-console-oauth-config\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.894725 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46knd\" (UniqueName: \"kubernetes.io/projected/36af9a72-1276-40cc-b462-84af24cf5abd-kube-api-access-46knd\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.894765 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36af9a72-1276-40cc-b462-84af24cf5abd-console-serving-cert\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.894854 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-service-ca\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.894894 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-console-config\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.894937 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-trusted-ca-bundle\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.996351 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46knd\" (UniqueName: \"kubernetes.io/projected/36af9a72-1276-40cc-b462-84af24cf5abd-kube-api-access-46knd\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.996662 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36af9a72-1276-40cc-b462-84af24cf5abd-console-serving-cert\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.996714 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-service-ca\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.996740 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-console-config\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.996762 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-trusted-ca-bundle\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.996812 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-oauth-serving-cert\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.996830 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36af9a72-1276-40cc-b462-84af24cf5abd-console-oauth-config\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:29 crc kubenswrapper[4715]: I1210 09:45:29.999751 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-service-ca\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.000474 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/36af9a72-1276-40cc-b462-84af24cf5abd-console-oauth-config\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.001203 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-oauth-serving-cert\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.001716 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-trusted-ca-bundle\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.001762 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/36af9a72-1276-40cc-b462-84af24cf5abd-console-config\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.006262 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/36af9a72-1276-40cc-b462-84af24cf5abd-console-serving-cert\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.018756 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46knd\" (UniqueName: \"kubernetes.io/projected/36af9a72-1276-40cc-b462-84af24cf5abd-kube-api-access-46knd\") pod \"console-6b86b8bf9d-69c54\" (UID: \"36af9a72-1276-40cc-b462-84af24cf5abd\") " pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.201482 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bc0de16a-d1b8-41b0-ad21-6015ea156e9a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-nsj47\" (UID: \"bc0de16a-d1b8-41b0-ad21-6015ea156e9a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.207013 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bc0de16a-d1b8-41b0-ad21-6015ea156e9a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-nsj47\" (UID: \"bc0de16a-d1b8-41b0-ad21-6015ea156e9a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.225798 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.303778 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2993b76a-4f33-4a02-8845-e694f2f65179-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-9srkc\" (UID: \"2993b76a-4f33-4a02-8845-e694f2f65179\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.304476 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z"] Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.308639 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2993b76a-4f33-4a02-8845-e694f2f65179-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-9srkc\" (UID: \"2993b76a-4f33-4a02-8845-e694f2f65179\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:30 crc kubenswrapper[4715]: W1210 09:45:30.308650 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadb59670_27dd_447f_8ccc_ca87956810ec.slice/crio-a96e06f3c8197a98eae3c282920648742ef5ee78d111904fd6f5c1b1cebc9db6 WatchSource:0}: Error finding container a96e06f3c8197a98eae3c282920648742ef5ee78d111904fd6f5c1b1cebc9db6: Status 404 returned error can't find the container with id a96e06f3c8197a98eae3c282920648742ef5ee78d111904fd6f5c1b1cebc9db6 Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.429265 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6b86b8bf9d-69c54"] Dec 10 09:45:30 crc kubenswrapper[4715]: W1210 09:45:30.442656 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36af9a72_1276_40cc_b462_84af24cf5abd.slice/crio-e8bc92a4ae65bd579633f1f9ffb13f0ef56422f201bde02476826b71d2c71362 WatchSource:0}: Error finding container e8bc92a4ae65bd579633f1f9ffb13f0ef56422f201bde02476826b71d2c71362: Status 404 returned error can't find the container with id e8bc92a4ae65bd579633f1f9ffb13f0ef56422f201bde02476826b71d2c71362 Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.463416 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.568812 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z" event={"ID":"adb59670-27dd-447f-8ccc-ca87956810ec","Type":"ContainerStarted","Data":"a96e06f3c8197a98eae3c282920648742ef5ee78d111904fd6f5c1b1cebc9db6"} Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.570150 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6b86b8bf9d-69c54" event={"ID":"36af9a72-1276-40cc-b462-84af24cf5abd","Type":"ContainerStarted","Data":"e8bc92a4ae65bd579633f1f9ffb13f0ef56422f201bde02476826b71d2c71362"} Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.574743 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hhxxc" event={"ID":"11a0c637-5b67-4db5-8fd0-33c6fabc6103","Type":"ContainerStarted","Data":"85f8f93014eb919359fcee40ad40f367bf744be7292a59f3d6ae7f5b2bc2afb0"} Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.576901 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.652531 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47"] Dec 10 09:45:30 crc kubenswrapper[4715]: W1210 09:45:30.661603 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc0de16a_d1b8_41b0_ad21_6015ea156e9a.slice/crio-3a415822f7a257383b94e893e3749711ba1818c1898e9ab3a952b9a907a85c14 WatchSource:0}: Error finding container 3a415822f7a257383b94e893e3749711ba1818c1898e9ab3a952b9a907a85c14: Status 404 returned error can't find the container with id 3a415822f7a257383b94e893e3749711ba1818c1898e9ab3a952b9a907a85c14 Dec 10 09:45:30 crc kubenswrapper[4715]: I1210 09:45:30.789088 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc"] Dec 10 09:45:30 crc kubenswrapper[4715]: W1210 09:45:30.792175 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2993b76a_4f33_4a02_8845_e694f2f65179.slice/crio-550ff645cb66a8636ccb4846fb6fa75feaf2e119df51d6cd6d6406d6f3ff5763 WatchSource:0}: Error finding container 550ff645cb66a8636ccb4846fb6fa75feaf2e119df51d6cd6d6406d6f3ff5763: Status 404 returned error can't find the container with id 550ff645cb66a8636ccb4846fb6fa75feaf2e119df51d6cd6d6406d6f3ff5763 Dec 10 09:45:31 crc kubenswrapper[4715]: I1210 09:45:31.583733 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6b86b8bf9d-69c54" event={"ID":"36af9a72-1276-40cc-b462-84af24cf5abd","Type":"ContainerStarted","Data":"1838050d2b77fde35ff5a84d4f6896b9829f5b4bf0f97b47016f362e89595ba4"} Dec 10 09:45:31 crc kubenswrapper[4715]: I1210 09:45:31.585317 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" event={"ID":"bc0de16a-d1b8-41b0-ad21-6015ea156e9a","Type":"ContainerStarted","Data":"3a415822f7a257383b94e893e3749711ba1818c1898e9ab3a952b9a907a85c14"} Dec 10 09:45:31 crc kubenswrapper[4715]: I1210 09:45:31.586723 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" event={"ID":"2993b76a-4f33-4a02-8845-e694f2f65179","Type":"ContainerStarted","Data":"550ff645cb66a8636ccb4846fb6fa75feaf2e119df51d6cd6d6406d6f3ff5763"} Dec 10 09:45:31 crc kubenswrapper[4715]: I1210 09:45:31.604647 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6b86b8bf9d-69c54" podStartSLOduration=2.604629264 podStartE2EDuration="2.604629264s" podCreationTimestamp="2025-12-10 09:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:45:31.601623671 +0000 UTC m=+694.345169942" watchObservedRunningTime="2025-12-10 09:45:31.604629264 +0000 UTC m=+694.348175515" Dec 10 09:45:35 crc kubenswrapper[4715]: I1210 09:45:35.644285 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hhxxc" event={"ID":"11a0c637-5b67-4db5-8fd0-33c6fabc6103","Type":"ContainerStarted","Data":"1eab21b821be8d4f20502ae44d5fa4b44050740a9cc4833adbbe8b1775f60b41"} Dec 10 09:45:35 crc kubenswrapper[4715]: I1210 09:45:35.646122 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:35 crc kubenswrapper[4715]: I1210 09:45:35.650070 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" event={"ID":"bc0de16a-d1b8-41b0-ad21-6015ea156e9a","Type":"ContainerStarted","Data":"38cec6a190bead0f6e20f08f7857ed668a71f349804b934b229f516776a9398d"} Dec 10 09:45:35 crc kubenswrapper[4715]: I1210 09:45:35.650200 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:45:35 crc kubenswrapper[4715]: I1210 09:45:35.725264 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-hhxxc" podStartSLOduration=1.53628587 podStartE2EDuration="6.725230402s" podCreationTimestamp="2025-12-10 09:45:29 +0000 UTC" firstStartedPulling="2025-12-10 09:45:29.95406168 +0000 UTC m=+692.697607931" lastFinishedPulling="2025-12-10 09:45:35.143006202 +0000 UTC m=+697.886552463" observedRunningTime="2025-12-10 09:45:35.724750918 +0000 UTC m=+698.468297169" watchObservedRunningTime="2025-12-10 09:45:35.725230402 +0000 UTC m=+698.468776653" Dec 10 09:45:35 crc kubenswrapper[4715]: I1210 09:45:35.751939 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" podStartSLOduration=2.263546932 podStartE2EDuration="6.75192213s" podCreationTimestamp="2025-12-10 09:45:29 +0000 UTC" firstStartedPulling="2025-12-10 09:45:30.663721185 +0000 UTC m=+693.407267436" lastFinishedPulling="2025-12-10 09:45:35.152096393 +0000 UTC m=+697.895642634" observedRunningTime="2025-12-10 09:45:35.748769873 +0000 UTC m=+698.492316124" watchObservedRunningTime="2025-12-10 09:45:35.75192213 +0000 UTC m=+698.495468381" Dec 10 09:45:36 crc kubenswrapper[4715]: I1210 09:45:36.656705 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z" event={"ID":"adb59670-27dd-447f-8ccc-ca87956810ec","Type":"ContainerStarted","Data":"d48d39744bc5f0bef19f09616fb9bb36ec20ca61adfde9b91d5ad300cccfae1d"} Dec 10 09:45:37 crc kubenswrapper[4715]: I1210 09:45:37.663740 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" event={"ID":"2993b76a-4f33-4a02-8845-e694f2f65179","Type":"ContainerStarted","Data":"89393aaae050c91fa016fba371e0a8802882f8574b244954153b07ad5e55fe25"} Dec 10 09:45:37 crc kubenswrapper[4715]: I1210 09:45:37.682029 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9srkc" podStartSLOduration=2.758653584 podStartE2EDuration="8.682000474s" podCreationTimestamp="2025-12-10 09:45:29 +0000 UTC" firstStartedPulling="2025-12-10 09:45:30.795439777 +0000 UTC m=+693.538986028" lastFinishedPulling="2025-12-10 09:45:36.718786667 +0000 UTC m=+699.462332918" observedRunningTime="2025-12-10 09:45:37.679558956 +0000 UTC m=+700.423105217" watchObservedRunningTime="2025-12-10 09:45:37.682000474 +0000 UTC m=+700.425546735" Dec 10 09:45:40 crc kubenswrapper[4715]: I1210 09:45:40.226051 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:40 crc kubenswrapper[4715]: I1210 09:45:40.227258 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:40 crc kubenswrapper[4715]: I1210 09:45:40.232549 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:40 crc kubenswrapper[4715]: I1210 09:45:40.684731 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6b86b8bf9d-69c54" Dec 10 09:45:40 crc kubenswrapper[4715]: I1210 09:45:40.744969 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zphsm"] Dec 10 09:45:41 crc kubenswrapper[4715]: I1210 09:45:41.688567 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z" event={"ID":"adb59670-27dd-447f-8ccc-ca87956810ec","Type":"ContainerStarted","Data":"652e8205d889e110a54b9003bb7cbb15cf0443754f70379c5555be216d01b551"} Dec 10 09:45:41 crc kubenswrapper[4715]: I1210 09:45:41.722442 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lwx7z" podStartSLOduration=2.195485169 podStartE2EDuration="12.722143007s" podCreationTimestamp="2025-12-10 09:45:29 +0000 UTC" firstStartedPulling="2025-12-10 09:45:30.31135331 +0000 UTC m=+693.054899561" lastFinishedPulling="2025-12-10 09:45:40.838011148 +0000 UTC m=+703.581557399" observedRunningTime="2025-12-10 09:45:41.706576567 +0000 UTC m=+704.450122848" watchObservedRunningTime="2025-12-10 09:45:41.722143007 +0000 UTC m=+704.465689268" Dec 10 09:45:44 crc kubenswrapper[4715]: I1210 09:45:44.902042 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-hhxxc" Dec 10 09:45:50 crc kubenswrapper[4715]: I1210 09:45:50.469530 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-nsj47" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.613740 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj"] Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.615676 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.617741 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.623761 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj"] Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.753074 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.753133 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.753205 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28zmq\" (UniqueName: \"kubernetes.io/projected/a71d4c3d-8768-4bed-a14b-cf226090ef7d-kube-api-access-28zmq\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.854670 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.854732 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.854770 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28zmq\" (UniqueName: \"kubernetes.io/projected/a71d4c3d-8768-4bed-a14b-cf226090ef7d-kube-api-access-28zmq\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.855183 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.855295 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.874224 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28zmq\" (UniqueName: \"kubernetes.io/projected/a71d4c3d-8768-4bed-a14b-cf226090ef7d-kube-api-access-28zmq\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:03 crc kubenswrapper[4715]: I1210 09:46:03.931863 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:04 crc kubenswrapper[4715]: I1210 09:46:04.111660 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj"] Dec 10 09:46:04 crc kubenswrapper[4715]: I1210 09:46:04.826088 4715 generic.go:334] "Generic (PLEG): container finished" podID="a71d4c3d-8768-4bed-a14b-cf226090ef7d" containerID="e8a7c4252551e8b75b637208e43d58f196f6c056dabb8c1441d56196f185547c" exitCode=0 Dec 10 09:46:04 crc kubenswrapper[4715]: I1210 09:46:04.826224 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" event={"ID":"a71d4c3d-8768-4bed-a14b-cf226090ef7d","Type":"ContainerDied","Data":"e8a7c4252551e8b75b637208e43d58f196f6c056dabb8c1441d56196f185547c"} Dec 10 09:46:04 crc kubenswrapper[4715]: I1210 09:46:04.826761 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" event={"ID":"a71d4c3d-8768-4bed-a14b-cf226090ef7d","Type":"ContainerStarted","Data":"014a35b7fe86539988bd6227fb739d49592f2a95bd57b5fccf3f02544ea62f61"} Dec 10 09:46:05 crc kubenswrapper[4715]: I1210 09:46:05.789341 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-zphsm" podUID="5fbc3706-b2a3-48aa-a950-dc403907fa02" containerName="console" containerID="cri-o://9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14" gracePeriod=15 Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.223977 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zphsm_5fbc3706-b2a3-48aa-a950-dc403907fa02/console/0.log" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.224309 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.283405 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-serving-cert\") pod \"5fbc3706-b2a3-48aa-a950-dc403907fa02\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.283442 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk5kx\" (UniqueName: \"kubernetes.io/projected/5fbc3706-b2a3-48aa-a950-dc403907fa02-kube-api-access-jk5kx\") pod \"5fbc3706-b2a3-48aa-a950-dc403907fa02\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.283473 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-config\") pod \"5fbc3706-b2a3-48aa-a950-dc403907fa02\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.283541 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-trusted-ca-bundle\") pod \"5fbc3706-b2a3-48aa-a950-dc403907fa02\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.283602 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-service-ca\") pod \"5fbc3706-b2a3-48aa-a950-dc403907fa02\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.283626 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-oauth-config\") pod \"5fbc3706-b2a3-48aa-a950-dc403907fa02\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.283642 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-oauth-serving-cert\") pod \"5fbc3706-b2a3-48aa-a950-dc403907fa02\" (UID: \"5fbc3706-b2a3-48aa-a950-dc403907fa02\") " Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.284359 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "5fbc3706-b2a3-48aa-a950-dc403907fa02" (UID: "5fbc3706-b2a3-48aa-a950-dc403907fa02"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.284423 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "5fbc3706-b2a3-48aa-a950-dc403907fa02" (UID: "5fbc3706-b2a3-48aa-a950-dc403907fa02"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.284532 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-service-ca" (OuterVolumeSpecName: "service-ca") pod "5fbc3706-b2a3-48aa-a950-dc403907fa02" (UID: "5fbc3706-b2a3-48aa-a950-dc403907fa02"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.284831 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-config" (OuterVolumeSpecName: "console-config") pod "5fbc3706-b2a3-48aa-a950-dc403907fa02" (UID: "5fbc3706-b2a3-48aa-a950-dc403907fa02"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.289285 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fbc3706-b2a3-48aa-a950-dc403907fa02-kube-api-access-jk5kx" (OuterVolumeSpecName: "kube-api-access-jk5kx") pod "5fbc3706-b2a3-48aa-a950-dc403907fa02" (UID: "5fbc3706-b2a3-48aa-a950-dc403907fa02"). InnerVolumeSpecName "kube-api-access-jk5kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.289556 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "5fbc3706-b2a3-48aa-a950-dc403907fa02" (UID: "5fbc3706-b2a3-48aa-a950-dc403907fa02"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.289684 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "5fbc3706-b2a3-48aa-a950-dc403907fa02" (UID: "5fbc3706-b2a3-48aa-a950-dc403907fa02"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.385527 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.385568 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.385577 4715 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.385590 4715 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.385599 4715 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.385607 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk5kx\" (UniqueName: \"kubernetes.io/projected/5fbc3706-b2a3-48aa-a950-dc403907fa02-kube-api-access-jk5kx\") on node \"crc\" DevicePath \"\"" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.385617 4715 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5fbc3706-b2a3-48aa-a950-dc403907fa02-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.839009 4715 generic.go:334] "Generic (PLEG): container finished" podID="a71d4c3d-8768-4bed-a14b-cf226090ef7d" containerID="543a6d742acdbb738fa4947aac7048ea0945a2bb0da4833cb5cd12e2377f14c9" exitCode=0 Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.839081 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" event={"ID":"a71d4c3d-8768-4bed-a14b-cf226090ef7d","Type":"ContainerDied","Data":"543a6d742acdbb738fa4947aac7048ea0945a2bb0da4833cb5cd12e2377f14c9"} Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.842025 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zphsm_5fbc3706-b2a3-48aa-a950-dc403907fa02/console/0.log" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.842070 4715 generic.go:334] "Generic (PLEG): container finished" podID="5fbc3706-b2a3-48aa-a950-dc403907fa02" containerID="9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14" exitCode=2 Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.842106 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zphsm" event={"ID":"5fbc3706-b2a3-48aa-a950-dc403907fa02","Type":"ContainerDied","Data":"9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14"} Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.842137 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zphsm" event={"ID":"5fbc3706-b2a3-48aa-a950-dc403907fa02","Type":"ContainerDied","Data":"053328968d72f359580dbec95909ef7d7ae9e5580ccf916ec1c5ec658a9e9f07"} Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.842163 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zphsm" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.842177 4715 scope.go:117] "RemoveContainer" containerID="9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.866862 4715 scope.go:117] "RemoveContainer" containerID="9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14" Dec 10 09:46:06 crc kubenswrapper[4715]: E1210 09:46:06.868578 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14\": container with ID starting with 9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14 not found: ID does not exist" containerID="9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.868612 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14"} err="failed to get container status \"9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14\": rpc error: code = NotFound desc = could not find container \"9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14\": container with ID starting with 9d30c06b5a92f840be8326332cede02d03a4128f3c65f19dcef8b2cc87ceda14 not found: ID does not exist" Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.886264 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zphsm"] Dec 10 09:46:06 crc kubenswrapper[4715]: I1210 09:46:06.889855 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-zphsm"] Dec 10 09:46:07 crc kubenswrapper[4715]: I1210 09:46:07.621655 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fbc3706-b2a3-48aa-a950-dc403907fa02" path="/var/lib/kubelet/pods/5fbc3706-b2a3-48aa-a950-dc403907fa02/volumes" Dec 10 09:46:07 crc kubenswrapper[4715]: I1210 09:46:07.853671 4715 generic.go:334] "Generic (PLEG): container finished" podID="a71d4c3d-8768-4bed-a14b-cf226090ef7d" containerID="214c0c6607c7da7407ab16876bcffb175261a6a8ae15dcd8a602155de9e6adc5" exitCode=0 Dec 10 09:46:07 crc kubenswrapper[4715]: I1210 09:46:07.853722 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" event={"ID":"a71d4c3d-8768-4bed-a14b-cf226090ef7d","Type":"ContainerDied","Data":"214c0c6607c7da7407ab16876bcffb175261a6a8ae15dcd8a602155de9e6adc5"} Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.073407 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.220997 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-util\") pod \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.221072 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-bundle\") pod \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.221223 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28zmq\" (UniqueName: \"kubernetes.io/projected/a71d4c3d-8768-4bed-a14b-cf226090ef7d-kube-api-access-28zmq\") pod \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\" (UID: \"a71d4c3d-8768-4bed-a14b-cf226090ef7d\") " Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.222290 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-bundle" (OuterVolumeSpecName: "bundle") pod "a71d4c3d-8768-4bed-a14b-cf226090ef7d" (UID: "a71d4c3d-8768-4bed-a14b-cf226090ef7d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.226242 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a71d4c3d-8768-4bed-a14b-cf226090ef7d-kube-api-access-28zmq" (OuterVolumeSpecName: "kube-api-access-28zmq") pod "a71d4c3d-8768-4bed-a14b-cf226090ef7d" (UID: "a71d4c3d-8768-4bed-a14b-cf226090ef7d"). InnerVolumeSpecName "kube-api-access-28zmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.236036 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-util" (OuterVolumeSpecName: "util") pod "a71d4c3d-8768-4bed-a14b-cf226090ef7d" (UID: "a71d4c3d-8768-4bed-a14b-cf226090ef7d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.322858 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-util\") on node \"crc\" DevicePath \"\"" Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.322897 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a71d4c3d-8768-4bed-a14b-cf226090ef7d-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.322929 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28zmq\" (UniqueName: \"kubernetes.io/projected/a71d4c3d-8768-4bed-a14b-cf226090ef7d-kube-api-access-28zmq\") on node \"crc\" DevicePath \"\"" Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.867459 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" event={"ID":"a71d4c3d-8768-4bed-a14b-cf226090ef7d","Type":"ContainerDied","Data":"014a35b7fe86539988bd6227fb739d49592f2a95bd57b5fccf3f02544ea62f61"} Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.867508 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="014a35b7fe86539988bd6227fb739d49592f2a95bd57b5fccf3f02544ea62f61" Dec 10 09:46:09 crc kubenswrapper[4715]: I1210 09:46:09.867512 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.191901 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv"] Dec 10 09:46:18 crc kubenswrapper[4715]: E1210 09:46:18.193343 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a71d4c3d-8768-4bed-a14b-cf226090ef7d" containerName="util" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.193420 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a71d4c3d-8768-4bed-a14b-cf226090ef7d" containerName="util" Dec 10 09:46:18 crc kubenswrapper[4715]: E1210 09:46:18.193487 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fbc3706-b2a3-48aa-a950-dc403907fa02" containerName="console" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.193534 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fbc3706-b2a3-48aa-a950-dc403907fa02" containerName="console" Dec 10 09:46:18 crc kubenswrapper[4715]: E1210 09:46:18.193589 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a71d4c3d-8768-4bed-a14b-cf226090ef7d" containerName="extract" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.193635 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a71d4c3d-8768-4bed-a14b-cf226090ef7d" containerName="extract" Dec 10 09:46:18 crc kubenswrapper[4715]: E1210 09:46:18.193683 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a71d4c3d-8768-4bed-a14b-cf226090ef7d" containerName="pull" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.193727 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a71d4c3d-8768-4bed-a14b-cf226090ef7d" containerName="pull" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.193896 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fbc3706-b2a3-48aa-a950-dc403907fa02" containerName="console" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.194019 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a71d4c3d-8768-4bed-a14b-cf226090ef7d" containerName="extract" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.194497 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.197141 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.197507 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.197729 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.198018 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.198271 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-jfxhd" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.241359 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b85b47f-761c-4064-8a3e-21824cb5eef7-webhook-cert\") pod \"metallb-operator-controller-manager-69456cd765-w9pmv\" (UID: \"6b85b47f-761c-4064-8a3e-21824cb5eef7\") " pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.241463 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6qfp\" (UniqueName: \"kubernetes.io/projected/6b85b47f-761c-4064-8a3e-21824cb5eef7-kube-api-access-v6qfp\") pod \"metallb-operator-controller-manager-69456cd765-w9pmv\" (UID: \"6b85b47f-761c-4064-8a3e-21824cb5eef7\") " pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.241566 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b85b47f-761c-4064-8a3e-21824cb5eef7-apiservice-cert\") pod \"metallb-operator-controller-manager-69456cd765-w9pmv\" (UID: \"6b85b47f-761c-4064-8a3e-21824cb5eef7\") " pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.249058 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv"] Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.342884 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b85b47f-761c-4064-8a3e-21824cb5eef7-webhook-cert\") pod \"metallb-operator-controller-manager-69456cd765-w9pmv\" (UID: \"6b85b47f-761c-4064-8a3e-21824cb5eef7\") " pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.342973 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6qfp\" (UniqueName: \"kubernetes.io/projected/6b85b47f-761c-4064-8a3e-21824cb5eef7-kube-api-access-v6qfp\") pod \"metallb-operator-controller-manager-69456cd765-w9pmv\" (UID: \"6b85b47f-761c-4064-8a3e-21824cb5eef7\") " pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.343069 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b85b47f-761c-4064-8a3e-21824cb5eef7-apiservice-cert\") pod \"metallb-operator-controller-manager-69456cd765-w9pmv\" (UID: \"6b85b47f-761c-4064-8a3e-21824cb5eef7\") " pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.349145 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b85b47f-761c-4064-8a3e-21824cb5eef7-webhook-cert\") pod \"metallb-operator-controller-manager-69456cd765-w9pmv\" (UID: \"6b85b47f-761c-4064-8a3e-21824cb5eef7\") " pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.349214 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b85b47f-761c-4064-8a3e-21824cb5eef7-apiservice-cert\") pod \"metallb-operator-controller-manager-69456cd765-w9pmv\" (UID: \"6b85b47f-761c-4064-8a3e-21824cb5eef7\") " pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.368443 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6qfp\" (UniqueName: \"kubernetes.io/projected/6b85b47f-761c-4064-8a3e-21824cb5eef7-kube-api-access-v6qfp\") pod \"metallb-operator-controller-manager-69456cd765-w9pmv\" (UID: \"6b85b47f-761c-4064-8a3e-21824cb5eef7\") " pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.439848 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv"] Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.440623 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.445298 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-bmxf8" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.449745 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.456058 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.457889 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv"] Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.512834 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.545240 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/711fc7a7-527a-4e3b-b343-a279ce6d3b79-webhook-cert\") pod \"metallb-operator-webhook-server-79dd69b788-gwkcv\" (UID: \"711fc7a7-527a-4e3b-b343-a279ce6d3b79\") " pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.545335 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/711fc7a7-527a-4e3b-b343-a279ce6d3b79-apiservice-cert\") pod \"metallb-operator-webhook-server-79dd69b788-gwkcv\" (UID: \"711fc7a7-527a-4e3b-b343-a279ce6d3b79\") " pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.545393 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6c2k\" (UniqueName: \"kubernetes.io/projected/711fc7a7-527a-4e3b-b343-a279ce6d3b79-kube-api-access-k6c2k\") pod \"metallb-operator-webhook-server-79dd69b788-gwkcv\" (UID: \"711fc7a7-527a-4e3b-b343-a279ce6d3b79\") " pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.646699 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6c2k\" (UniqueName: \"kubernetes.io/projected/711fc7a7-527a-4e3b-b343-a279ce6d3b79-kube-api-access-k6c2k\") pod \"metallb-operator-webhook-server-79dd69b788-gwkcv\" (UID: \"711fc7a7-527a-4e3b-b343-a279ce6d3b79\") " pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.646771 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/711fc7a7-527a-4e3b-b343-a279ce6d3b79-webhook-cert\") pod \"metallb-operator-webhook-server-79dd69b788-gwkcv\" (UID: \"711fc7a7-527a-4e3b-b343-a279ce6d3b79\") " pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.646867 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/711fc7a7-527a-4e3b-b343-a279ce6d3b79-apiservice-cert\") pod \"metallb-operator-webhook-server-79dd69b788-gwkcv\" (UID: \"711fc7a7-527a-4e3b-b343-a279ce6d3b79\") " pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.651041 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/711fc7a7-527a-4e3b-b343-a279ce6d3b79-apiservice-cert\") pod \"metallb-operator-webhook-server-79dd69b788-gwkcv\" (UID: \"711fc7a7-527a-4e3b-b343-a279ce6d3b79\") " pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.651744 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/711fc7a7-527a-4e3b-b343-a279ce6d3b79-webhook-cert\") pod \"metallb-operator-webhook-server-79dd69b788-gwkcv\" (UID: \"711fc7a7-527a-4e3b-b343-a279ce6d3b79\") " pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.670723 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6c2k\" (UniqueName: \"kubernetes.io/projected/711fc7a7-527a-4e3b-b343-a279ce6d3b79-kube-api-access-k6c2k\") pod \"metallb-operator-webhook-server-79dd69b788-gwkcv\" (UID: \"711fc7a7-527a-4e3b-b343-a279ce6d3b79\") " pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.762225 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:18 crc kubenswrapper[4715]: I1210 09:46:18.941080 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv"] Dec 10 09:46:18 crc kubenswrapper[4715]: W1210 09:46:18.948155 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b85b47f_761c_4064_8a3e_21824cb5eef7.slice/crio-3aa3b70dd2b8789544e5a58da947be4b1060679c8d077b679edc85832d3afbe6 WatchSource:0}: Error finding container 3aa3b70dd2b8789544e5a58da947be4b1060679c8d077b679edc85832d3afbe6: Status 404 returned error can't find the container with id 3aa3b70dd2b8789544e5a58da947be4b1060679c8d077b679edc85832d3afbe6 Dec 10 09:46:19 crc kubenswrapper[4715]: I1210 09:46:19.058194 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv"] Dec 10 09:46:19 crc kubenswrapper[4715]: W1210 09:46:19.063774 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod711fc7a7_527a_4e3b_b343_a279ce6d3b79.slice/crio-89355fbd4e2e88cd599044ee63d5ce9ca9f8d0a6848da5e1239c29a573cc8d0a WatchSource:0}: Error finding container 89355fbd4e2e88cd599044ee63d5ce9ca9f8d0a6848da5e1239c29a573cc8d0a: Status 404 returned error can't find the container with id 89355fbd4e2e88cd599044ee63d5ce9ca9f8d0a6848da5e1239c29a573cc8d0a Dec 10 09:46:19 crc kubenswrapper[4715]: I1210 09:46:19.931611 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" event={"ID":"711fc7a7-527a-4e3b-b343-a279ce6d3b79","Type":"ContainerStarted","Data":"89355fbd4e2e88cd599044ee63d5ce9ca9f8d0a6848da5e1239c29a573cc8d0a"} Dec 10 09:46:19 crc kubenswrapper[4715]: I1210 09:46:19.932848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" event={"ID":"6b85b47f-761c-4064-8a3e-21824cb5eef7","Type":"ContainerStarted","Data":"3aa3b70dd2b8789544e5a58da947be4b1060679c8d077b679edc85832d3afbe6"} Dec 10 09:46:27 crc kubenswrapper[4715]: I1210 09:46:27.980845 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" event={"ID":"711fc7a7-527a-4e3b-b343-a279ce6d3b79","Type":"ContainerStarted","Data":"f22510e8a041f24c7c56b38383afd1ef2bd2a994e50fe4fab22a55eae9e21bd2"} Dec 10 09:46:27 crc kubenswrapper[4715]: I1210 09:46:27.982142 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:27 crc kubenswrapper[4715]: I1210 09:46:27.984208 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" event={"ID":"6b85b47f-761c-4064-8a3e-21824cb5eef7","Type":"ContainerStarted","Data":"242d27f1171fa61fa69566f5b7e47e9ebde4a121fc92f85e55a6ebee6fa5ee45"} Dec 10 09:46:27 crc kubenswrapper[4715]: I1210 09:46:27.984566 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:28 crc kubenswrapper[4715]: I1210 09:46:28.039436 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" podStartSLOduration=1.488466493 podStartE2EDuration="10.039413872s" podCreationTimestamp="2025-12-10 09:46:18 +0000 UTC" firstStartedPulling="2025-12-10 09:46:19.068603527 +0000 UTC m=+741.812149778" lastFinishedPulling="2025-12-10 09:46:27.619550906 +0000 UTC m=+750.363097157" observedRunningTime="2025-12-10 09:46:28.00336803 +0000 UTC m=+750.746914291" watchObservedRunningTime="2025-12-10 09:46:28.039413872 +0000 UTC m=+750.782960133" Dec 10 09:46:38 crc kubenswrapper[4715]: I1210 09:46:38.768520 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-79dd69b788-gwkcv" Dec 10 09:46:38 crc kubenswrapper[4715]: I1210 09:46:38.793697 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" podStartSLOduration=12.145143549 podStartE2EDuration="20.793678176s" podCreationTimestamp="2025-12-10 09:46:18 +0000 UTC" firstStartedPulling="2025-12-10 09:46:18.952620858 +0000 UTC m=+741.696167109" lastFinishedPulling="2025-12-10 09:46:27.601155485 +0000 UTC m=+750.344701736" observedRunningTime="2025-12-10 09:46:28.03757354 +0000 UTC m=+750.781119791" watchObservedRunningTime="2025-12-10 09:46:38.793678176 +0000 UTC m=+761.537224437" Dec 10 09:46:42 crc kubenswrapper[4715]: I1210 09:46:42.414304 4715 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 09:46:47 crc kubenswrapper[4715]: I1210 09:46:47.714150 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:46:47 crc kubenswrapper[4715]: I1210 09:46:47.714690 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:46:58 crc kubenswrapper[4715]: I1210 09:46:58.516334 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-69456cd765-w9pmv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.279258 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg"] Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.280297 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.286891 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-p2fz4"] Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.287732 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-mq58j" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.288031 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.292089 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.296178 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.296362 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.297013 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg"] Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.345260 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slqb5\" (UniqueName: \"kubernetes.io/projected/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-kube-api-access-slqb5\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.345305 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-metrics\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.345327 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-frr-conf\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.345407 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7f61745-e2a2-4e9a-b2a1-b931599251a8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-s5rwg\" (UID: \"e7f61745-e2a2-4e9a-b2a1-b931599251a8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.345440 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-reloader\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.345459 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-frr-startup\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.345492 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8ft9\" (UniqueName: \"kubernetes.io/projected/e7f61745-e2a2-4e9a-b2a1-b931599251a8-kube-api-access-r8ft9\") pod \"frr-k8s-webhook-server-7fcb986d4-s5rwg\" (UID: \"e7f61745-e2a2-4e9a-b2a1-b931599251a8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.345601 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-metrics-certs\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.345655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-frr-sockets\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.368062 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-g6jgv"] Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.369312 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.371802 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.372395 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-xsr9b"] Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.373017 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.375261 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-hrkwh" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.376162 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.376616 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.378445 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.390171 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-xsr9b"] Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447136 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7f61745-e2a2-4e9a-b2a1-b931599251a8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-s5rwg\" (UID: \"e7f61745-e2a2-4e9a-b2a1-b931599251a8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447178 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6f9bf056-705e-4f97-a470-1bd98ae14e30-metallb-excludel2\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447209 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-reloader\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447223 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-frr-startup\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447245 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-metrics-certs\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447270 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgljf\" (UniqueName: \"kubernetes.io/projected/fd252c98-3f30-4b3f-aa36-4056f00035a2-kube-api-access-sgljf\") pod \"controller-f8648f98b-xsr9b\" (UID: \"fd252c98-3f30-4b3f-aa36-4056f00035a2\") " pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447291 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8ft9\" (UniqueName: \"kubernetes.io/projected/e7f61745-e2a2-4e9a-b2a1-b931599251a8-kube-api-access-r8ft9\") pod \"frr-k8s-webhook-server-7fcb986d4-s5rwg\" (UID: \"e7f61745-e2a2-4e9a-b2a1-b931599251a8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-metrics-certs\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447329 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-frr-sockets\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: E1210 09:46:59.447327 4715 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447350 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm85s\" (UniqueName: \"kubernetes.io/projected/6f9bf056-705e-4f97-a470-1bd98ae14e30-kube-api-access-dm85s\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447367 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fd252c98-3f30-4b3f-aa36-4056f00035a2-metrics-certs\") pod \"controller-f8648f98b-xsr9b\" (UID: \"fd252c98-3f30-4b3f-aa36-4056f00035a2\") " pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: E1210 09:46:59.447412 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7f61745-e2a2-4e9a-b2a1-b931599251a8-cert podName:e7f61745-e2a2-4e9a-b2a1-b931599251a8 nodeName:}" failed. No retries permitted until 2025-12-10 09:46:59.94739384 +0000 UTC m=+782.690940091 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e7f61745-e2a2-4e9a-b2a1-b931599251a8-cert") pod "frr-k8s-webhook-server-7fcb986d4-s5rwg" (UID: "e7f61745-e2a2-4e9a-b2a1-b931599251a8") : secret "frr-k8s-webhook-server-cert" not found Dec 10 09:46:59 crc kubenswrapper[4715]: E1210 09:46:59.447505 4715 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447530 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slqb5\" (UniqueName: \"kubernetes.io/projected/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-kube-api-access-slqb5\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: E1210 09:46:59.447593 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-metrics-certs podName:0fd6b12e-bdcd-4662-b1d4-5b2605b42295 nodeName:}" failed. No retries permitted until 2025-12-10 09:46:59.947567075 +0000 UTC m=+782.691113386 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-metrics-certs") pod "frr-k8s-p2fz4" (UID: "0fd6b12e-bdcd-4662-b1d4-5b2605b42295") : secret "frr-k8s-certs-secret" not found Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447630 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-metrics\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447703 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-frr-conf\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447769 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-memberlist\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447831 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd252c98-3f30-4b3f-aa36-4056f00035a2-cert\") pod \"controller-f8648f98b-xsr9b\" (UID: \"fd252c98-3f30-4b3f-aa36-4056f00035a2\") " pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447768 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-reloader\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.447860 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-frr-sockets\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.448033 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-frr-conf\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.448151 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-metrics\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.449162 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-frr-startup\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.473447 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slqb5\" (UniqueName: \"kubernetes.io/projected/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-kube-api-access-slqb5\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.480613 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8ft9\" (UniqueName: \"kubernetes.io/projected/e7f61745-e2a2-4e9a-b2a1-b931599251a8-kube-api-access-r8ft9\") pod \"frr-k8s-webhook-server-7fcb986d4-s5rwg\" (UID: \"e7f61745-e2a2-4e9a-b2a1-b931599251a8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.549621 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgljf\" (UniqueName: \"kubernetes.io/projected/fd252c98-3f30-4b3f-aa36-4056f00035a2-kube-api-access-sgljf\") pod \"controller-f8648f98b-xsr9b\" (UID: \"fd252c98-3f30-4b3f-aa36-4056f00035a2\") " pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.549698 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm85s\" (UniqueName: \"kubernetes.io/projected/6f9bf056-705e-4f97-a470-1bd98ae14e30-kube-api-access-dm85s\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.549721 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fd252c98-3f30-4b3f-aa36-4056f00035a2-metrics-certs\") pod \"controller-f8648f98b-xsr9b\" (UID: \"fd252c98-3f30-4b3f-aa36-4056f00035a2\") " pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.549754 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-memberlist\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.549778 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd252c98-3f30-4b3f-aa36-4056f00035a2-cert\") pod \"controller-f8648f98b-xsr9b\" (UID: \"fd252c98-3f30-4b3f-aa36-4056f00035a2\") " pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.549817 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6f9bf056-705e-4f97-a470-1bd98ae14e30-metallb-excludel2\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.549847 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-metrics-certs\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: E1210 09:46:59.549972 4715 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 10 09:46:59 crc kubenswrapper[4715]: E1210 09:46:59.550022 4715 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 10 09:46:59 crc kubenswrapper[4715]: E1210 09:46:59.550069 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-memberlist podName:6f9bf056-705e-4f97-a470-1bd98ae14e30 nodeName:}" failed. No retries permitted until 2025-12-10 09:47:00.050045395 +0000 UTC m=+782.793591646 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-memberlist") pod "speaker-g6jgv" (UID: "6f9bf056-705e-4f97-a470-1bd98ae14e30") : secret "metallb-memberlist" not found Dec 10 09:46:59 crc kubenswrapper[4715]: E1210 09:46:59.550091 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-metrics-certs podName:6f9bf056-705e-4f97-a470-1bd98ae14e30 nodeName:}" failed. No retries permitted until 2025-12-10 09:47:00.050082746 +0000 UTC m=+782.793628997 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-metrics-certs") pod "speaker-g6jgv" (UID: "6f9bf056-705e-4f97-a470-1bd98ae14e30") : secret "speaker-certs-secret" not found Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.550814 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6f9bf056-705e-4f97-a470-1bd98ae14e30-metallb-excludel2\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.551448 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.563779 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd252c98-3f30-4b3f-aa36-4056f00035a2-cert\") pod \"controller-f8648f98b-xsr9b\" (UID: \"fd252c98-3f30-4b3f-aa36-4056f00035a2\") " pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.566281 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm85s\" (UniqueName: \"kubernetes.io/projected/6f9bf056-705e-4f97-a470-1bd98ae14e30-kube-api-access-dm85s\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.566608 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fd252c98-3f30-4b3f-aa36-4056f00035a2-metrics-certs\") pod \"controller-f8648f98b-xsr9b\" (UID: \"fd252c98-3f30-4b3f-aa36-4056f00035a2\") " pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.571411 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgljf\" (UniqueName: \"kubernetes.io/projected/fd252c98-3f30-4b3f-aa36-4056f00035a2-kube-api-access-sgljf\") pod \"controller-f8648f98b-xsr9b\" (UID: \"fd252c98-3f30-4b3f-aa36-4056f00035a2\") " pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.692126 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.928300 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-xsr9b"] Dec 10 09:46:59 crc kubenswrapper[4715]: W1210 09:46:59.931779 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd252c98_3f30_4b3f_aa36_4056f00035a2.slice/crio-6b150b9616fac25952d329bfcd7c921ed34f83bb11c4f53722212ea29a293218 WatchSource:0}: Error finding container 6b150b9616fac25952d329bfcd7c921ed34f83bb11c4f53722212ea29a293218: Status 404 returned error can't find the container with id 6b150b9616fac25952d329bfcd7c921ed34f83bb11c4f53722212ea29a293218 Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.954534 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7f61745-e2a2-4e9a-b2a1-b931599251a8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-s5rwg\" (UID: \"e7f61745-e2a2-4e9a-b2a1-b931599251a8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.954684 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-metrics-certs\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.959810 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0fd6b12e-bdcd-4662-b1d4-5b2605b42295-metrics-certs\") pod \"frr-k8s-p2fz4\" (UID: \"0fd6b12e-bdcd-4662-b1d4-5b2605b42295\") " pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:46:59 crc kubenswrapper[4715]: I1210 09:46:59.959879 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7f61745-e2a2-4e9a-b2a1-b931599251a8-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-s5rwg\" (UID: \"e7f61745-e2a2-4e9a-b2a1-b931599251a8\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:47:00 crc kubenswrapper[4715]: I1210 09:47:00.056428 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-memberlist\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:47:00 crc kubenswrapper[4715]: I1210 09:47:00.056508 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-metrics-certs\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:47:00 crc kubenswrapper[4715]: E1210 09:47:00.056634 4715 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 10 09:47:00 crc kubenswrapper[4715]: E1210 09:47:00.056714 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-memberlist podName:6f9bf056-705e-4f97-a470-1bd98ae14e30 nodeName:}" failed. No retries permitted until 2025-12-10 09:47:01.056694668 +0000 UTC m=+783.800240919 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-memberlist") pod "speaker-g6jgv" (UID: "6f9bf056-705e-4f97-a470-1bd98ae14e30") : secret "metallb-memberlist" not found Dec 10 09:47:00 crc kubenswrapper[4715]: I1210 09:47:00.059900 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-metrics-certs\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:47:00 crc kubenswrapper[4715]: I1210 09:47:00.194944 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-xsr9b" event={"ID":"fd252c98-3f30-4b3f-aa36-4056f00035a2","Type":"ContainerStarted","Data":"7f7f786621670f84313b0b9bc03b3ebd2debc33aab235391ab1e64783257cfb7"} Dec 10 09:47:00 crc kubenswrapper[4715]: I1210 09:47:00.195008 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-xsr9b" event={"ID":"fd252c98-3f30-4b3f-aa36-4056f00035a2","Type":"ContainerStarted","Data":"6b150b9616fac25952d329bfcd7c921ed34f83bb11c4f53722212ea29a293218"} Dec 10 09:47:00 crc kubenswrapper[4715]: I1210 09:47:00.197219 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:47:00 crc kubenswrapper[4715]: I1210 09:47:00.216314 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:47:00 crc kubenswrapper[4715]: I1210 09:47:00.444818 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg"] Dec 10 09:47:00 crc kubenswrapper[4715]: W1210 09:47:00.454482 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7f61745_e2a2_4e9a_b2a1_b931599251a8.slice/crio-531fed287606be6c67019b4b7c15afa4e5fb9ed180b0274865238456f2f9c9b8 WatchSource:0}: Error finding container 531fed287606be6c67019b4b7c15afa4e5fb9ed180b0274865238456f2f9c9b8: Status 404 returned error can't find the container with id 531fed287606be6c67019b4b7c15afa4e5fb9ed180b0274865238456f2f9c9b8 Dec 10 09:47:01 crc kubenswrapper[4715]: I1210 09:47:01.070890 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-memberlist\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:47:01 crc kubenswrapper[4715]: I1210 09:47:01.089137 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f9bf056-705e-4f97-a470-1bd98ae14e30-memberlist\") pod \"speaker-g6jgv\" (UID: \"6f9bf056-705e-4f97-a470-1bd98ae14e30\") " pod="metallb-system/speaker-g6jgv" Dec 10 09:47:01 crc kubenswrapper[4715]: I1210 09:47:01.188830 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-g6jgv" Dec 10 09:47:01 crc kubenswrapper[4715]: I1210 09:47:01.201073 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" event={"ID":"e7f61745-e2a2-4e9a-b2a1-b931599251a8","Type":"ContainerStarted","Data":"531fed287606be6c67019b4b7c15afa4e5fb9ed180b0274865238456f2f9c9b8"} Dec 10 09:47:01 crc kubenswrapper[4715]: I1210 09:47:01.203052 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-xsr9b" event={"ID":"fd252c98-3f30-4b3f-aa36-4056f00035a2","Type":"ContainerStarted","Data":"f834a5f5deae55a973864a518b787c7c8809f2a3fad3c320cf80a4a1dd8dd6c0"} Dec 10 09:47:01 crc kubenswrapper[4715]: I1210 09:47:01.203209 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:47:01 crc kubenswrapper[4715]: I1210 09:47:01.204990 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p2fz4" event={"ID":"0fd6b12e-bdcd-4662-b1d4-5b2605b42295","Type":"ContainerStarted","Data":"d1589774e5d5011377b856fc6dd4d9fe9ceb91f8c47a83214e6c5175473b3dd1"} Dec 10 09:47:01 crc kubenswrapper[4715]: I1210 09:47:01.222605 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-xsr9b" podStartSLOduration=2.222591377 podStartE2EDuration="2.222591377s" podCreationTimestamp="2025-12-10 09:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:47:01.220473837 +0000 UTC m=+783.964020088" watchObservedRunningTime="2025-12-10 09:47:01.222591377 +0000 UTC m=+783.966137628" Dec 10 09:47:02 crc kubenswrapper[4715]: I1210 09:47:02.216012 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-g6jgv" event={"ID":"6f9bf056-705e-4f97-a470-1bd98ae14e30","Type":"ContainerStarted","Data":"aa75fdeee34a82c10f3d72498d96cd8daf129126fc6c0c3bbd3224a44044e3a1"} Dec 10 09:47:02 crc kubenswrapper[4715]: I1210 09:47:02.216358 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-g6jgv" event={"ID":"6f9bf056-705e-4f97-a470-1bd98ae14e30","Type":"ContainerStarted","Data":"a1ac6d3e0f197e6b35d186e801474032160cd5ba80bdc5dca7fb7f10dcab5032"} Dec 10 09:47:02 crc kubenswrapper[4715]: I1210 09:47:02.216378 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-g6jgv" event={"ID":"6f9bf056-705e-4f97-a470-1bd98ae14e30","Type":"ContainerStarted","Data":"0daacc62b4a098c06f40358cc7a9fb701f33f2070d94eeef5dde6689b3268ea1"} Dec 10 09:47:02 crc kubenswrapper[4715]: I1210 09:47:02.216547 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-g6jgv" Dec 10 09:47:02 crc kubenswrapper[4715]: I1210 09:47:02.243179 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-g6jgv" podStartSLOduration=3.243162129 podStartE2EDuration="3.243162129s" podCreationTimestamp="2025-12-10 09:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:47:02.242402577 +0000 UTC m=+784.985948828" watchObservedRunningTime="2025-12-10 09:47:02.243162129 +0000 UTC m=+784.986708380" Dec 10 09:47:09 crc kubenswrapper[4715]: I1210 09:47:09.305036 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" event={"ID":"e7f61745-e2a2-4e9a-b2a1-b931599251a8","Type":"ContainerStarted","Data":"79bd33922e0de2cdbaa61fe89fc3de250896592f33d3e35580d90bdf8150dfa5"} Dec 10 09:47:09 crc kubenswrapper[4715]: I1210 09:47:09.305533 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:47:09 crc kubenswrapper[4715]: I1210 09:47:09.307887 4715 generic.go:334] "Generic (PLEG): container finished" podID="0fd6b12e-bdcd-4662-b1d4-5b2605b42295" containerID="ea1bcab8e56bbeaaae622eddfed65313a61980c2a5a5058f8a2c9be043b2a30a" exitCode=0 Dec 10 09:47:09 crc kubenswrapper[4715]: I1210 09:47:09.307945 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p2fz4" event={"ID":"0fd6b12e-bdcd-4662-b1d4-5b2605b42295","Type":"ContainerDied","Data":"ea1bcab8e56bbeaaae622eddfed65313a61980c2a5a5058f8a2c9be043b2a30a"} Dec 10 09:47:09 crc kubenswrapper[4715]: I1210 09:47:09.324247 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" podStartSLOduration=2.581356412 podStartE2EDuration="10.324229773s" podCreationTimestamp="2025-12-10 09:46:59 +0000 UTC" firstStartedPulling="2025-12-10 09:47:00.461239313 +0000 UTC m=+783.204785564" lastFinishedPulling="2025-12-10 09:47:08.204112674 +0000 UTC m=+790.947658925" observedRunningTime="2025-12-10 09:47:09.318425728 +0000 UTC m=+792.061971989" watchObservedRunningTime="2025-12-10 09:47:09.324229773 +0000 UTC m=+792.067776024" Dec 10 09:47:10 crc kubenswrapper[4715]: I1210 09:47:10.316259 4715 generic.go:334] "Generic (PLEG): container finished" podID="0fd6b12e-bdcd-4662-b1d4-5b2605b42295" containerID="6aee42a50e6ca967070130a4668ca8d5e734d8fb318aad9ac34dc607a355b12c" exitCode=0 Dec 10 09:47:10 crc kubenswrapper[4715]: I1210 09:47:10.316329 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p2fz4" event={"ID":"0fd6b12e-bdcd-4662-b1d4-5b2605b42295","Type":"ContainerDied","Data":"6aee42a50e6ca967070130a4668ca8d5e734d8fb318aad9ac34dc607a355b12c"} Dec 10 09:47:11 crc kubenswrapper[4715]: I1210 09:47:11.195937 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-g6jgv" Dec 10 09:47:11 crc kubenswrapper[4715]: I1210 09:47:11.332012 4715 generic.go:334] "Generic (PLEG): container finished" podID="0fd6b12e-bdcd-4662-b1d4-5b2605b42295" containerID="36ccb2db18beed21819512b2391675d7989bbc809da2890797332f124bdaefac" exitCode=0 Dec 10 09:47:11 crc kubenswrapper[4715]: I1210 09:47:11.332067 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p2fz4" event={"ID":"0fd6b12e-bdcd-4662-b1d4-5b2605b42295","Type":"ContainerDied","Data":"36ccb2db18beed21819512b2391675d7989bbc809da2890797332f124bdaefac"} Dec 10 09:47:12 crc kubenswrapper[4715]: I1210 09:47:12.345559 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p2fz4" event={"ID":"0fd6b12e-bdcd-4662-b1d4-5b2605b42295","Type":"ContainerStarted","Data":"f168fa87c3c5f394f6760bdde8e81fd55ba78e4805dea80638683f649dada42f"} Dec 10 09:47:12 crc kubenswrapper[4715]: I1210 09:47:12.345873 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p2fz4" event={"ID":"0fd6b12e-bdcd-4662-b1d4-5b2605b42295","Type":"ContainerStarted","Data":"02c64b878e081ec4938fb6f4dce5465d114a3ee7b13e32fd9e4917e2236d19e9"} Dec 10 09:47:12 crc kubenswrapper[4715]: I1210 09:47:12.345886 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p2fz4" event={"ID":"0fd6b12e-bdcd-4662-b1d4-5b2605b42295","Type":"ContainerStarted","Data":"7dab6ba30127a42c0c2d303bb31a17b4ad37a94b5b1b7ac4ad343f2619b1dfc1"} Dec 10 09:47:12 crc kubenswrapper[4715]: I1210 09:47:12.345897 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p2fz4" event={"ID":"0fd6b12e-bdcd-4662-b1d4-5b2605b42295","Type":"ContainerStarted","Data":"547a8f1a57db6bb02b0ba29aa97e5622ee1eab7d7c3117a398ddd1f646476154"} Dec 10 09:47:12 crc kubenswrapper[4715]: I1210 09:47:12.345923 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p2fz4" event={"ID":"0fd6b12e-bdcd-4662-b1d4-5b2605b42295","Type":"ContainerStarted","Data":"4597464a69426baa8a27602450a3d29deb918eb077af2a6786b439cf0c45a31a"} Dec 10 09:47:13 crc kubenswrapper[4715]: I1210 09:47:13.359164 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-p2fz4" event={"ID":"0fd6b12e-bdcd-4662-b1d4-5b2605b42295","Type":"ContainerStarted","Data":"eeb7894823952a7b1393095dc82c76d85d9007928973d3cb4c98bc5b92da3001"} Dec 10 09:47:13 crc kubenswrapper[4715]: I1210 09:47:13.359365 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:47:13 crc kubenswrapper[4715]: I1210 09:47:13.389087 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-p2fz4" podStartSLOduration=6.466345941 podStartE2EDuration="14.389061349s" podCreationTimestamp="2025-12-10 09:46:59 +0000 UTC" firstStartedPulling="2025-12-10 09:47:00.301908129 +0000 UTC m=+783.045454380" lastFinishedPulling="2025-12-10 09:47:08.224623537 +0000 UTC m=+790.968169788" observedRunningTime="2025-12-10 09:47:13.383788919 +0000 UTC m=+796.127335170" watchObservedRunningTime="2025-12-10 09:47:13.389061349 +0000 UTC m=+796.132607600" Dec 10 09:47:15 crc kubenswrapper[4715]: I1210 09:47:15.217346 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:47:15 crc kubenswrapper[4715]: I1210 09:47:15.265116 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.664298 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-crsjq"] Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.665569 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-crsjq" Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.668212 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-bmg87" Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.668774 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.668867 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.671498 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-crsjq"] Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.714183 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.714485 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.784895 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np9r4\" (UniqueName: \"kubernetes.io/projected/313f8227-63e5-435e-a2e9-0c08d67712b8-kube-api-access-np9r4\") pod \"openstack-operator-index-crsjq\" (UID: \"313f8227-63e5-435e-a2e9-0c08d67712b8\") " pod="openstack-operators/openstack-operator-index-crsjq" Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.886367 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np9r4\" (UniqueName: \"kubernetes.io/projected/313f8227-63e5-435e-a2e9-0c08d67712b8-kube-api-access-np9r4\") pod \"openstack-operator-index-crsjq\" (UID: \"313f8227-63e5-435e-a2e9-0c08d67712b8\") " pod="openstack-operators/openstack-operator-index-crsjq" Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.907587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np9r4\" (UniqueName: \"kubernetes.io/projected/313f8227-63e5-435e-a2e9-0c08d67712b8-kube-api-access-np9r4\") pod \"openstack-operator-index-crsjq\" (UID: \"313f8227-63e5-435e-a2e9-0c08d67712b8\") " pod="openstack-operators/openstack-operator-index-crsjq" Dec 10 09:47:17 crc kubenswrapper[4715]: I1210 09:47:17.991217 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-crsjq" Dec 10 09:47:18 crc kubenswrapper[4715]: I1210 09:47:18.390900 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-crsjq"] Dec 10 09:47:18 crc kubenswrapper[4715]: W1210 09:47:18.406122 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod313f8227_63e5_435e_a2e9_0c08d67712b8.slice/crio-122a1e06b450aa9dd431383822996ac7f1aa2a3fb6d97221c9353613af4ae381 WatchSource:0}: Error finding container 122a1e06b450aa9dd431383822996ac7f1aa2a3fb6d97221c9353613af4ae381: Status 404 returned error can't find the container with id 122a1e06b450aa9dd431383822996ac7f1aa2a3fb6d97221c9353613af4ae381 Dec 10 09:47:18 crc kubenswrapper[4715]: I1210 09:47:18.431205 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-crsjq" event={"ID":"313f8227-63e5-435e-a2e9-0c08d67712b8","Type":"ContainerStarted","Data":"122a1e06b450aa9dd431383822996ac7f1aa2a3fb6d97221c9353613af4ae381"} Dec 10 09:47:19 crc kubenswrapper[4715]: I1210 09:47:19.696224 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-xsr9b" Dec 10 09:47:20 crc kubenswrapper[4715]: I1210 09:47:20.204016 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-s5rwg" Dec 10 09:47:22 crc kubenswrapper[4715]: I1210 09:47:22.468751 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-crsjq" event={"ID":"313f8227-63e5-435e-a2e9-0c08d67712b8","Type":"ContainerStarted","Data":"821e378f8b0c94d02be5a0059aeede906edeaa72cf5c121558437970c9faaaa1"} Dec 10 09:47:22 crc kubenswrapper[4715]: I1210 09:47:22.482652 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-crsjq" podStartSLOduration=1.873455194 podStartE2EDuration="5.482631045s" podCreationTimestamp="2025-12-10 09:47:17 +0000 UTC" firstStartedPulling="2025-12-10 09:47:18.40762042 +0000 UTC m=+801.151166671" lastFinishedPulling="2025-12-10 09:47:22.016796271 +0000 UTC m=+804.760342522" observedRunningTime="2025-12-10 09:47:22.480590777 +0000 UTC m=+805.224137028" watchObservedRunningTime="2025-12-10 09:47:22.482631045 +0000 UTC m=+805.226177296" Dec 10 09:47:27 crc kubenswrapper[4715]: I1210 09:47:27.992116 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-crsjq" Dec 10 09:47:27 crc kubenswrapper[4715]: I1210 09:47:27.992454 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-crsjq" Dec 10 09:47:28 crc kubenswrapper[4715]: I1210 09:47:28.028681 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-crsjq" Dec 10 09:47:28 crc kubenswrapper[4715]: I1210 09:47:28.529400 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-crsjq" Dec 10 09:47:30 crc kubenswrapper[4715]: I1210 09:47:30.295582 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-p2fz4" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.117405 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk"] Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.118898 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.126012 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk"] Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.127085 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6vflm" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.310455 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-bundle\") pod \"9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.310563 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-util\") pod \"9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.310592 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftk44\" (UniqueName: \"kubernetes.io/projected/c077e8ab-c825-493d-b731-b3cc36eac159-kube-api-access-ftk44\") pod \"9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.411492 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-util\") pod \"9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.411558 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftk44\" (UniqueName: \"kubernetes.io/projected/c077e8ab-c825-493d-b731-b3cc36eac159-kube-api-access-ftk44\") pod \"9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.411612 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-bundle\") pod \"9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.412046 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-util\") pod \"9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.412125 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-bundle\") pod \"9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.428942 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftk44\" (UniqueName: \"kubernetes.io/projected/c077e8ab-c825-493d-b731-b3cc36eac159-kube-api-access-ftk44\") pod \"9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.434603 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:31 crc kubenswrapper[4715]: I1210 09:47:31.970751 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk"] Dec 10 09:47:31 crc kubenswrapper[4715]: W1210 09:47:31.983015 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc077e8ab_c825_493d_b731_b3cc36eac159.slice/crio-34e31daa07716f81b95cd3dfcbfa093bbc8fc08f3763945382e10127564736fd WatchSource:0}: Error finding container 34e31daa07716f81b95cd3dfcbfa093bbc8fc08f3763945382e10127564736fd: Status 404 returned error can't find the container with id 34e31daa07716f81b95cd3dfcbfa093bbc8fc08f3763945382e10127564736fd Dec 10 09:47:32 crc kubenswrapper[4715]: I1210 09:47:32.525388 4715 generic.go:334] "Generic (PLEG): container finished" podID="c077e8ab-c825-493d-b731-b3cc36eac159" containerID="ff341a4c56d7133647d47ebffdee78ac23dbf276a524f10fabbbc861c40dcf44" exitCode=0 Dec 10 09:47:32 crc kubenswrapper[4715]: I1210 09:47:32.525459 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" event={"ID":"c077e8ab-c825-493d-b731-b3cc36eac159","Type":"ContainerDied","Data":"ff341a4c56d7133647d47ebffdee78ac23dbf276a524f10fabbbc861c40dcf44"} Dec 10 09:47:32 crc kubenswrapper[4715]: I1210 09:47:32.525685 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" event={"ID":"c077e8ab-c825-493d-b731-b3cc36eac159","Type":"ContainerStarted","Data":"34e31daa07716f81b95cd3dfcbfa093bbc8fc08f3763945382e10127564736fd"} Dec 10 09:47:34 crc kubenswrapper[4715]: I1210 09:47:34.539457 4715 generic.go:334] "Generic (PLEG): container finished" podID="c077e8ab-c825-493d-b731-b3cc36eac159" containerID="fb5015cb193f384f2541253d08428719f33396f9bbcdc88cbc98b6c0760a76a1" exitCode=0 Dec 10 09:47:34 crc kubenswrapper[4715]: I1210 09:47:34.539504 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" event={"ID":"c077e8ab-c825-493d-b731-b3cc36eac159","Type":"ContainerDied","Data":"fb5015cb193f384f2541253d08428719f33396f9bbcdc88cbc98b6c0760a76a1"} Dec 10 09:47:35 crc kubenswrapper[4715]: I1210 09:47:35.549258 4715 generic.go:334] "Generic (PLEG): container finished" podID="c077e8ab-c825-493d-b731-b3cc36eac159" containerID="f956ffb4680371e9ae0ecd8fefbff7a7d57e28e41fcc69c452cad52618c0d0aa" exitCode=0 Dec 10 09:47:35 crc kubenswrapper[4715]: I1210 09:47:35.549521 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" event={"ID":"c077e8ab-c825-493d-b731-b3cc36eac159","Type":"ContainerDied","Data":"f956ffb4680371e9ae0ecd8fefbff7a7d57e28e41fcc69c452cad52618c0d0aa"} Dec 10 09:47:36 crc kubenswrapper[4715]: I1210 09:47:36.791584 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:36 crc kubenswrapper[4715]: I1210 09:47:36.861983 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-bundle\") pod \"c077e8ab-c825-493d-b731-b3cc36eac159\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " Dec 10 09:47:36 crc kubenswrapper[4715]: I1210 09:47:36.862047 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-util\") pod \"c077e8ab-c825-493d-b731-b3cc36eac159\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " Dec 10 09:47:36 crc kubenswrapper[4715]: I1210 09:47:36.862176 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftk44\" (UniqueName: \"kubernetes.io/projected/c077e8ab-c825-493d-b731-b3cc36eac159-kube-api-access-ftk44\") pod \"c077e8ab-c825-493d-b731-b3cc36eac159\" (UID: \"c077e8ab-c825-493d-b731-b3cc36eac159\") " Dec 10 09:47:36 crc kubenswrapper[4715]: I1210 09:47:36.862564 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-bundle" (OuterVolumeSpecName: "bundle") pod "c077e8ab-c825-493d-b731-b3cc36eac159" (UID: "c077e8ab-c825-493d-b731-b3cc36eac159"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:47:36 crc kubenswrapper[4715]: I1210 09:47:36.867719 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c077e8ab-c825-493d-b731-b3cc36eac159-kube-api-access-ftk44" (OuterVolumeSpecName: "kube-api-access-ftk44") pod "c077e8ab-c825-493d-b731-b3cc36eac159" (UID: "c077e8ab-c825-493d-b731-b3cc36eac159"). InnerVolumeSpecName "kube-api-access-ftk44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:47:36 crc kubenswrapper[4715]: I1210 09:47:36.878634 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-util" (OuterVolumeSpecName: "util") pod "c077e8ab-c825-493d-b731-b3cc36eac159" (UID: "c077e8ab-c825-493d-b731-b3cc36eac159"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:47:36 crc kubenswrapper[4715]: I1210 09:47:36.964718 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:47:36 crc kubenswrapper[4715]: I1210 09:47:36.964774 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c077e8ab-c825-493d-b731-b3cc36eac159-util\") on node \"crc\" DevicePath \"\"" Dec 10 09:47:36 crc kubenswrapper[4715]: I1210 09:47:36.964793 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftk44\" (UniqueName: \"kubernetes.io/projected/c077e8ab-c825-493d-b731-b3cc36eac159-kube-api-access-ftk44\") on node \"crc\" DevicePath \"\"" Dec 10 09:47:37 crc kubenswrapper[4715]: I1210 09:47:37.571410 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" event={"ID":"c077e8ab-c825-493d-b731-b3cc36eac159","Type":"ContainerDied","Data":"34e31daa07716f81b95cd3dfcbfa093bbc8fc08f3763945382e10127564736fd"} Dec 10 09:47:37 crc kubenswrapper[4715]: I1210 09:47:37.571449 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34e31daa07716f81b95cd3dfcbfa093bbc8fc08f3763945382e10127564736fd" Dec 10 09:47:37 crc kubenswrapper[4715]: I1210 09:47:37.571473 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.436326 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw"] Dec 10 09:47:40 crc kubenswrapper[4715]: E1210 09:47:40.438195 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c077e8ab-c825-493d-b731-b3cc36eac159" containerName="util" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.438223 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c077e8ab-c825-493d-b731-b3cc36eac159" containerName="util" Dec 10 09:47:40 crc kubenswrapper[4715]: E1210 09:47:40.438237 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c077e8ab-c825-493d-b731-b3cc36eac159" containerName="pull" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.438247 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c077e8ab-c825-493d-b731-b3cc36eac159" containerName="pull" Dec 10 09:47:40 crc kubenswrapper[4715]: E1210 09:47:40.438258 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c077e8ab-c825-493d-b731-b3cc36eac159" containerName="extract" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.438269 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c077e8ab-c825-493d-b731-b3cc36eac159" containerName="extract" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.438441 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c077e8ab-c825-493d-b731-b3cc36eac159" containerName="extract" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.439017 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.441423 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-tpg28" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.525465 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw"] Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.548545 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42g64\" (UniqueName: \"kubernetes.io/projected/5899529c-0209-4c8f-8b67-37c9d08bbcf8-kube-api-access-42g64\") pod \"openstack-operator-controller-operator-7dcf9bf79d-7ftkw\" (UID: \"5899529c-0209-4c8f-8b67-37c9d08bbcf8\") " pod="openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.649902 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42g64\" (UniqueName: \"kubernetes.io/projected/5899529c-0209-4c8f-8b67-37c9d08bbcf8-kube-api-access-42g64\") pod \"openstack-operator-controller-operator-7dcf9bf79d-7ftkw\" (UID: \"5899529c-0209-4c8f-8b67-37c9d08bbcf8\") " pod="openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.669972 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42g64\" (UniqueName: \"kubernetes.io/projected/5899529c-0209-4c8f-8b67-37c9d08bbcf8-kube-api-access-42g64\") pod \"openstack-operator-controller-operator-7dcf9bf79d-7ftkw\" (UID: \"5899529c-0209-4c8f-8b67-37c9d08bbcf8\") " pod="openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw" Dec 10 09:47:40 crc kubenswrapper[4715]: I1210 09:47:40.769270 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw" Dec 10 09:47:41 crc kubenswrapper[4715]: I1210 09:47:41.145052 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw"] Dec 10 09:47:41 crc kubenswrapper[4715]: I1210 09:47:41.597416 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw" event={"ID":"5899529c-0209-4c8f-8b67-37c9d08bbcf8","Type":"ContainerStarted","Data":"22cb875baed00c8e5ee160d58d27adbb0cba1bf851c7663ec15e35289ab65c19"} Dec 10 09:47:46 crc kubenswrapper[4715]: I1210 09:47:46.639353 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw" event={"ID":"5899529c-0209-4c8f-8b67-37c9d08bbcf8","Type":"ContainerStarted","Data":"dde17f951ca2bf61f6638062e3fe2607ca66ebb9cf558dcf344b0325cf7a9a7e"} Dec 10 09:47:46 crc kubenswrapper[4715]: I1210 09:47:46.640193 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw" Dec 10 09:47:47 crc kubenswrapper[4715]: I1210 09:47:47.713665 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:47:47 crc kubenswrapper[4715]: I1210 09:47:47.713974 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:47:47 crc kubenswrapper[4715]: I1210 09:47:47.714018 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:47:47 crc kubenswrapper[4715]: I1210 09:47:47.714659 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d20ad65a2bc63bec106125489b213c4e46e5f275365a8160d4b290979ccc044"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 09:47:47 crc kubenswrapper[4715]: I1210 09:47:47.714725 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://5d20ad65a2bc63bec106125489b213c4e46e5f275365a8160d4b290979ccc044" gracePeriod=600 Dec 10 09:47:48 crc kubenswrapper[4715]: I1210 09:47:48.658479 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="5d20ad65a2bc63bec106125489b213c4e46e5f275365a8160d4b290979ccc044" exitCode=0 Dec 10 09:47:48 crc kubenswrapper[4715]: I1210 09:47:48.658544 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"5d20ad65a2bc63bec106125489b213c4e46e5f275365a8160d4b290979ccc044"} Dec 10 09:47:48 crc kubenswrapper[4715]: I1210 09:47:48.659223 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"92c7512a624555c7386239654aadd7a7479838a440573c30b4fffcacf96a04bf"} Dec 10 09:47:48 crc kubenswrapper[4715]: I1210 09:47:48.659263 4715 scope.go:117] "RemoveContainer" containerID="ed1ab5cf91267ff02613fe03163e2a0e1e6fe3ab62fe2bfb0b6ed4dcd711b4fe" Dec 10 09:47:48 crc kubenswrapper[4715]: I1210 09:47:48.680434 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw" podStartSLOduration=3.656758655 podStartE2EDuration="8.680413321s" podCreationTimestamp="2025-12-10 09:47:40 +0000 UTC" firstStartedPulling="2025-12-10 09:47:41.153633334 +0000 UTC m=+823.897179585" lastFinishedPulling="2025-12-10 09:47:46.177288 +0000 UTC m=+828.920834251" observedRunningTime="2025-12-10 09:47:46.67221206 +0000 UTC m=+829.415758311" watchObservedRunningTime="2025-12-10 09:47:48.680413321 +0000 UTC m=+831.423959582" Dec 10 09:48:00 crc kubenswrapper[4715]: I1210 09:48:00.771801 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7dcf9bf79d-7ftkw" Dec 10 09:48:18 crc kubenswrapper[4715]: E1210 09:48:18.142479 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/systemd-tmpfiles-clean.service\": RecentStats: unable to find data in memory cache]" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.287381 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.288768 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.290466 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-jzqjr" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.301611 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.302607 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.307431 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.308334 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-vhnsf" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.327147 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.328770 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.332198 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-h94rb" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.347624 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.362655 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.397718 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.399863 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.429725 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-94s6k" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.430258 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzjsl\" (UniqueName: \"kubernetes.io/projected/af548670-106c-4fbc-b6a5-102896191190-kube-api-access-zzjsl\") pod \"cinder-operator-controller-manager-6c677c69b-htklg\" (UID: \"af548670-106c-4fbc-b6a5-102896191190\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.430353 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz8nk\" (UniqueName: \"kubernetes.io/projected/5949f2c9-f510-47f4-988c-cae06576536a-kube-api-access-fz8nk\") pod \"designate-operator-controller-manager-697fb699cf-j2drs\" (UID: \"5949f2c9-f510-47f4-988c-cae06576536a\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.430495 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf98q\" (UniqueName: \"kubernetes.io/projected/f1e4cba4-0694-4477-b031-5caac1fbcb84-kube-api-access-xf98q\") pod \"barbican-operator-controller-manager-7d9dfd778-jwwm5\" (UID: \"f1e4cba4-0694-4477-b031-5caac1fbcb84\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.440386 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.482217 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.483573 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.488326 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-mxkk6" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.490636 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.491805 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.496372 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-rjmvg" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.505516 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.506565 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.512145 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-xwjtw" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.512326 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.532687 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz8nk\" (UniqueName: \"kubernetes.io/projected/5949f2c9-f510-47f4-988c-cae06576536a-kube-api-access-fz8nk\") pod \"designate-operator-controller-manager-697fb699cf-j2drs\" (UID: \"5949f2c9-f510-47f4-988c-cae06576536a\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.532758 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9288\" (UniqueName: \"kubernetes.io/projected/06228f2c-8f0c-493a-850f-b1e5fd41a23b-kube-api-access-m9288\") pod \"horizon-operator-controller-manager-68c6d99b8f-d875q\" (UID: \"06228f2c-8f0c-493a-850f-b1e5fd41a23b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.532817 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvj66\" (UniqueName: \"kubernetes.io/projected/6c14bbb7-a3cf-4654-ba2c-2d993022fa72-kube-api-access-zvj66\") pod \"glance-operator-controller-manager-5697bb5779-scbxh\" (UID: \"6c14bbb7-a3cf-4654-ba2c-2d993022fa72\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.532834 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqrl6\" (UniqueName: \"kubernetes.io/projected/c396c828-63cf-4cd7-a050-c359ce8e1c8b-kube-api-access-fqrl6\") pod \"heat-operator-controller-manager-5f64f6f8bb-jdcr5\" (UID: \"c396c828-63cf-4cd7-a050-c359ce8e1c8b\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.532865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf98q\" (UniqueName: \"kubernetes.io/projected/f1e4cba4-0694-4477-b031-5caac1fbcb84-kube-api-access-xf98q\") pod \"barbican-operator-controller-manager-7d9dfd778-jwwm5\" (UID: \"f1e4cba4-0694-4477-b031-5caac1fbcb84\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.532906 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzjsl\" (UniqueName: \"kubernetes.io/projected/af548670-106c-4fbc-b6a5-102896191190-kube-api-access-zzjsl\") pod \"cinder-operator-controller-manager-6c677c69b-htklg\" (UID: \"af548670-106c-4fbc-b6a5-102896191190\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.538011 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.553843 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.554929 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.559656 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-v2jvw" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.560354 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.561756 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.568640 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-btp2q" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.569519 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.570832 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzjsl\" (UniqueName: \"kubernetes.io/projected/af548670-106c-4fbc-b6a5-102896191190-kube-api-access-zzjsl\") pod \"cinder-operator-controller-manager-6c677c69b-htklg\" (UID: \"af548670-106c-4fbc-b6a5-102896191190\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.571403 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz8nk\" (UniqueName: \"kubernetes.io/projected/5949f2c9-f510-47f4-988c-cae06576536a-kube-api-access-fz8nk\") pod \"designate-operator-controller-manager-697fb699cf-j2drs\" (UID: \"5949f2c9-f510-47f4-988c-cae06576536a\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.578642 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.580568 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf98q\" (UniqueName: \"kubernetes.io/projected/f1e4cba4-0694-4477-b031-5caac1fbcb84-kube-api-access-xf98q\") pod \"barbican-operator-controller-manager-7d9dfd778-jwwm5\" (UID: \"f1e4cba4-0694-4477-b031-5caac1fbcb84\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.590998 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.596994 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.605637 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.607275 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.610485 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-m2zqq" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.612528 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.616983 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.617352 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.620856 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.627113 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.631214 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-g5srx" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.634438 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvj66\" (UniqueName: \"kubernetes.io/projected/6c14bbb7-a3cf-4654-ba2c-2d993022fa72-kube-api-access-zvj66\") pod \"glance-operator-controller-manager-5697bb5779-scbxh\" (UID: \"6c14bbb7-a3cf-4654-ba2c-2d993022fa72\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.634477 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqrl6\" (UniqueName: \"kubernetes.io/projected/c396c828-63cf-4cd7-a050-c359ce8e1c8b-kube-api-access-fqrl6\") pod \"heat-operator-controller-manager-5f64f6f8bb-jdcr5\" (UID: \"c396c828-63cf-4cd7-a050-c359ce8e1c8b\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.634511 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2k75\" (UniqueName: \"kubernetes.io/projected/75ebabe2-8b57-4eee-9b3b-b238fafd94b7-kube-api-access-m2k75\") pod \"keystone-operator-controller-manager-7765d96ddf-88z54\" (UID: \"75ebabe2-8b57-4eee-9b3b-b238fafd94b7\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.634594 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbknf\" (UniqueName: \"kubernetes.io/projected/31619ce3-146d-4efb-898f-75bd1e007de9-kube-api-access-nbknf\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.634617 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m67tk\" (UniqueName: \"kubernetes.io/projected/c3397611-a5cb-455c-8c5e-dd24ab61545a-kube-api-access-m67tk\") pod \"ironic-operator-controller-manager-967d97867-h9rpn\" (UID: \"c3397611-a5cb-455c-8c5e-dd24ab61545a\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.634656 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.634698 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9288\" (UniqueName: \"kubernetes.io/projected/06228f2c-8f0c-493a-850f-b1e5fd41a23b-kube-api-access-m9288\") pod \"horizon-operator-controller-manager-68c6d99b8f-d875q\" (UID: \"06228f2c-8f0c-493a-850f-b1e5fd41a23b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.643679 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.672033 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.673104 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.674970 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.683841 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqrl6\" (UniqueName: \"kubernetes.io/projected/c396c828-63cf-4cd7-a050-c359ce8e1c8b-kube-api-access-fqrl6\") pod \"heat-operator-controller-manager-5f64f6f8bb-jdcr5\" (UID: \"c396c828-63cf-4cd7-a050-c359ce8e1c8b\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.684034 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvj66\" (UniqueName: \"kubernetes.io/projected/6c14bbb7-a3cf-4654-ba2c-2d993022fa72-kube-api-access-zvj66\") pod \"glance-operator-controller-manager-5697bb5779-scbxh\" (UID: \"6c14bbb7-a3cf-4654-ba2c-2d993022fa72\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.684729 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-mhndv" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.689574 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.691107 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9288\" (UniqueName: \"kubernetes.io/projected/06228f2c-8f0c-493a-850f-b1e5fd41a23b-kube-api-access-m9288\") pod \"horizon-operator-controller-manager-68c6d99b8f-d875q\" (UID: \"06228f2c-8f0c-493a-850f-b1e5fd41a23b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.691597 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.694225 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-7zl5m" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.702866 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.726143 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.734220 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.735808 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2k75\" (UniqueName: \"kubernetes.io/projected/75ebabe2-8b57-4eee-9b3b-b238fafd94b7-kube-api-access-m2k75\") pod \"keystone-operator-controller-manager-7765d96ddf-88z54\" (UID: \"75ebabe2-8b57-4eee-9b3b-b238fafd94b7\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.743092 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdgx7\" (UniqueName: \"kubernetes.io/projected/50f2991b-7b13-4d84-8a4d-504d9b873023-kube-api-access-cdgx7\") pod \"nova-operator-controller-manager-697bc559fc-f8zw7\" (UID: \"50f2991b-7b13-4d84-8a4d-504d9b873023\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.743580 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbknf\" (UniqueName: \"kubernetes.io/projected/31619ce3-146d-4efb-898f-75bd1e007de9-kube-api-access-nbknf\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.743828 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m67tk\" (UniqueName: \"kubernetes.io/projected/c3397611-a5cb-455c-8c5e-dd24ab61545a-kube-api-access-m67tk\") pod \"ironic-operator-controller-manager-967d97867-h9rpn\" (UID: \"c3397611-a5cb-455c-8c5e-dd24ab61545a\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.743995 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxjlq\" (UniqueName: \"kubernetes.io/projected/78ba49c9-f451-4982-8d88-6bf0444c8b60-kube-api-access-wxjlq\") pod \"mariadb-operator-controller-manager-79c8c4686c-tjqbt\" (UID: \"78ba49c9-f451-4982-8d88-6bf0444c8b60\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.744154 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.744305 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g5fz\" (UniqueName: \"kubernetes.io/projected/20025d92-9314-436f-a05e-4bfee3d3add0-kube-api-access-6g5fz\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-z25gz\" (UID: \"20025d92-9314-436f-a05e-4bfee3d3add0\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.741031 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" Dec 10 09:48:20 crc kubenswrapper[4715]: E1210 09:48:20.744475 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 09:48:20 crc kubenswrapper[4715]: E1210 09:48:20.746966 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert podName:31619ce3-146d-4efb-898f-75bd1e007de9 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:21.246941574 +0000 UTC m=+863.990487825 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert") pod "infra-operator-controller-manager-78d48bff9d-mxxcb" (UID: "31619ce3-146d-4efb-898f-75bd1e007de9") : secret "infra-operator-webhook-server-cert" not found Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.746255 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.750106 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swfpw\" (UniqueName: \"kubernetes.io/projected/d80992c0-9c12-4266-b705-6eaed7e46d1e-kube-api-access-swfpw\") pod \"manila-operator-controller-manager-5b5fd79c9c-p69tc\" (UID: \"d80992c0-9c12-4266-b705-6eaed7e46d1e\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.751871 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-9tfpg" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.764039 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.764090 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.764417 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.765181 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.767926 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.769217 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-t8j5h" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.769462 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-c26k5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.769885 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.776384 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.780889 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2k75\" (UniqueName: \"kubernetes.io/projected/75ebabe2-8b57-4eee-9b3b-b238fafd94b7-kube-api-access-m2k75\") pod \"keystone-operator-controller-manager-7765d96ddf-88z54\" (UID: \"75ebabe2-8b57-4eee-9b3b-b238fafd94b7\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.781592 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbknf\" (UniqueName: \"kubernetes.io/projected/31619ce3-146d-4efb-898f-75bd1e007de9-kube-api-access-nbknf\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.782158 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m67tk\" (UniqueName: \"kubernetes.io/projected/c3397611-a5cb-455c-8c5e-dd24ab61545a-kube-api-access-m67tk\") pod \"ironic-operator-controller-manager-967d97867-h9rpn\" (UID: \"c3397611-a5cb-455c-8c5e-dd24ab61545a\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.819298 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-m75xm"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.824311 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.844659 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.850548 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.852177 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxjlq\" (UniqueName: \"kubernetes.io/projected/78ba49c9-f451-4982-8d88-6bf0444c8b60-kube-api-access-wxjlq\") pod \"mariadb-operator-controller-manager-79c8c4686c-tjqbt\" (UID: \"78ba49c9-f451-4982-8d88-6bf0444c8b60\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.861173 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-m75xm"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.861841 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-fvhh8" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.875789 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g5fz\" (UniqueName: \"kubernetes.io/projected/20025d92-9314-436f-a05e-4bfee3d3add0-kube-api-access-6g5fz\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-z25gz\" (UID: \"20025d92-9314-436f-a05e-4bfee3d3add0\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.875855 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swfpw\" (UniqueName: \"kubernetes.io/projected/d80992c0-9c12-4266-b705-6eaed7e46d1e-kube-api-access-swfpw\") pod \"manila-operator-controller-manager-5b5fd79c9c-p69tc\" (UID: \"d80992c0-9c12-4266-b705-6eaed7e46d1e\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.875999 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcj94\" (UniqueName: \"kubernetes.io/projected/91566e3d-fbed-450a-bc66-1bf34f98d57f-kube-api-access-lcj94\") pod \"octavia-operator-controller-manager-998648c74-gbrf4\" (UID: \"91566e3d-fbed-450a-bc66-1bf34f98d57f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.876026 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b5qg\" (UniqueName: \"kubernetes.io/projected/01645280-6af5-4f8d-b412-3d7e0efa9054-kube-api-access-8b5qg\") pod \"ovn-operator-controller-manager-b6456fdb6-66fzd\" (UID: \"01645280-6af5-4f8d-b412-3d7e0efa9054\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.876118 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdgx7\" (UniqueName: \"kubernetes.io/projected/50f2991b-7b13-4d84-8a4d-504d9b873023-kube-api-access-cdgx7\") pod \"nova-operator-controller-manager-697bc559fc-f8zw7\" (UID: \"50f2991b-7b13-4d84-8a4d-504d9b873023\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.876145 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8llrh\" (UniqueName: \"kubernetes.io/projected/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-kube-api-access-8llrh\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.876169 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.919686 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxjlq\" (UniqueName: \"kubernetes.io/projected/78ba49c9-f451-4982-8d88-6bf0444c8b60-kube-api-access-wxjlq\") pod \"mariadb-operator-controller-manager-79c8c4686c-tjqbt\" (UID: \"78ba49c9-f451-4982-8d88-6bf0444c8b60\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.924451 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.925345 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swfpw\" (UniqueName: \"kubernetes.io/projected/d80992c0-9c12-4266-b705-6eaed7e46d1e-kube-api-access-swfpw\") pod \"manila-operator-controller-manager-5b5fd79c9c-p69tc\" (UID: \"d80992c0-9c12-4266-b705-6eaed7e46d1e\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.925950 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.927113 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.933264 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.934649 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-7sbxq" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.939957 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdgx7\" (UniqueName: \"kubernetes.io/projected/50f2991b-7b13-4d84-8a4d-504d9b873023-kube-api-access-cdgx7\") pod \"nova-operator-controller-manager-697bc559fc-f8zw7\" (UID: \"50f2991b-7b13-4d84-8a4d-504d9b873023\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.944999 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4"] Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.948181 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.948175 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g5fz\" (UniqueName: \"kubernetes.io/projected/20025d92-9314-436f-a05e-4bfee3d3add0-kube-api-access-6g5fz\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-z25gz\" (UID: \"20025d92-9314-436f-a05e-4bfee3d3add0\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" Dec 10 09:48:20 crc kubenswrapper[4715]: I1210 09:48:20.962281 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.079803 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcj94\" (UniqueName: \"kubernetes.io/projected/91566e3d-fbed-450a-bc66-1bf34f98d57f-kube-api-access-lcj94\") pod \"octavia-operator-controller-manager-998648c74-gbrf4\" (UID: \"91566e3d-fbed-450a-bc66-1bf34f98d57f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.079861 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnwfc\" (UniqueName: \"kubernetes.io/projected/629d31d3-4474-4beb-9c7b-2d890d2134e6-kube-api-access-qnwfc\") pod \"placement-operator-controller-manager-78f8948974-m75xm\" (UID: \"629d31d3-4474-4beb-9c7b-2d890d2134e6\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.079885 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b5qg\" (UniqueName: \"kubernetes.io/projected/01645280-6af5-4f8d-b412-3d7e0efa9054-kube-api-access-8b5qg\") pod \"ovn-operator-controller-manager-b6456fdb6-66fzd\" (UID: \"01645280-6af5-4f8d-b412-3d7e0efa9054\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.079943 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8llrh\" (UniqueName: \"kubernetes.io/projected/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-kube-api-access-8llrh\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.080314 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.080876 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" Dec 10 09:48:21 crc kubenswrapper[4715]: E1210 09:48:21.081559 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 09:48:21 crc kubenswrapper[4715]: E1210 09:48:21.081730 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert podName:d329770a-7d39-41d0-b4b1-a52e2e9ccda7 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:21.581704468 +0000 UTC m=+864.325250719 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fphth6" (UID: "d329770a-7d39-41d0-b4b1-a52e2e9ccda7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.084443 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.088174 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-4pn86" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.090053 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.090689 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.108713 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b5qg\" (UniqueName: \"kubernetes.io/projected/01645280-6af5-4f8d-b412-3d7e0efa9054-kube-api-access-8b5qg\") pod \"ovn-operator-controller-manager-b6456fdb6-66fzd\" (UID: \"01645280-6af5-4f8d-b412-3d7e0efa9054\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.109553 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.114677 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.115675 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcj94\" (UniqueName: \"kubernetes.io/projected/91566e3d-fbed-450a-bc66-1bf34f98d57f-kube-api-access-lcj94\") pod \"octavia-operator-controller-manager-998648c74-gbrf4\" (UID: \"91566e3d-fbed-450a-bc66-1bf34f98d57f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.119767 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8llrh\" (UniqueName: \"kubernetes.io/projected/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-kube-api-access-8llrh\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.146375 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.147558 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.150558 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-fq754" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.152805 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.175258 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.182060 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnwfc\" (UniqueName: \"kubernetes.io/projected/629d31d3-4474-4beb-9c7b-2d890d2134e6-kube-api-access-qnwfc\") pod \"placement-operator-controller-manager-78f8948974-m75xm\" (UID: \"629d31d3-4474-4beb-9c7b-2d890d2134e6\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.182116 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djjjd\" (UniqueName: \"kubernetes.io/projected/63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd-kube-api-access-djjjd\") pod \"telemetry-operator-controller-manager-58d5ff84df-2957k\" (UID: \"63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.182135 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z772\" (UniqueName: \"kubernetes.io/projected/cfe88ca5-7d1d-4fff-8404-8b192ae48b09-kube-api-access-2z772\") pod \"test-operator-controller-manager-5854674fcc-wcvvl\" (UID: \"cfe88ca5-7d1d-4fff-8404-8b192ae48b09\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.182160 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt6zc\" (UniqueName: \"kubernetes.io/projected/67b3cab4-c74e-4861-875b-e38aac7ced71-kube-api-access-rt6zc\") pod \"swift-operator-controller-manager-9d58d64bc-ht6n4\" (UID: \"67b3cab4-c74e-4861-875b-e38aac7ced71\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.213489 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.214699 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.216782 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-79pcl" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.217898 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.220426 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.239620 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.240541 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.243708 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.243731 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-mmtz5" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.243814 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.283539 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djjjd\" (UniqueName: \"kubernetes.io/projected/63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd-kube-api-access-djjjd\") pod \"telemetry-operator-controller-manager-58d5ff84df-2957k\" (UID: \"63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.283592 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z772\" (UniqueName: \"kubernetes.io/projected/cfe88ca5-7d1d-4fff-8404-8b192ae48b09-kube-api-access-2z772\") pod \"test-operator-controller-manager-5854674fcc-wcvvl\" (UID: \"cfe88ca5-7d1d-4fff-8404-8b192ae48b09\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.283634 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt6zc\" (UniqueName: \"kubernetes.io/projected/67b3cab4-c74e-4861-875b-e38aac7ced71-kube-api-access-rt6zc\") pod \"swift-operator-controller-manager-9d58d64bc-ht6n4\" (UID: \"67b3cab4-c74e-4861-875b-e38aac7ced71\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.283711 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:21 crc kubenswrapper[4715]: E1210 09:48:21.283887 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 09:48:21 crc kubenswrapper[4715]: E1210 09:48:21.283958 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert podName:31619ce3-146d-4efb-898f-75bd1e007de9 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:22.283940549 +0000 UTC m=+865.027486800 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert") pod "infra-operator-controller-manager-78d48bff9d-mxxcb" (UID: "31619ce3-146d-4efb-898f-75bd1e007de9") : secret "infra-operator-webhook-server-cert" not found Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.341279 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djjjd\" (UniqueName: \"kubernetes.io/projected/63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd-kube-api-access-djjjd\") pod \"telemetry-operator-controller-manager-58d5ff84df-2957k\" (UID: \"63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.348127 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt6zc\" (UniqueName: \"kubernetes.io/projected/67b3cab4-c74e-4861-875b-e38aac7ced71-kube-api-access-rt6zc\") pod \"swift-operator-controller-manager-9d58d64bc-ht6n4\" (UID: \"67b3cab4-c74e-4861-875b-e38aac7ced71\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.358504 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z772\" (UniqueName: \"kubernetes.io/projected/cfe88ca5-7d1d-4fff-8404-8b192ae48b09-kube-api-access-2z772\") pod \"test-operator-controller-manager-5854674fcc-wcvvl\" (UID: \"cfe88ca5-7d1d-4fff-8404-8b192ae48b09\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.358835 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnwfc\" (UniqueName: \"kubernetes.io/projected/629d31d3-4474-4beb-9c7b-2d890d2134e6-kube-api-access-qnwfc\") pod \"placement-operator-controller-manager-78f8948974-m75xm\" (UID: \"629d31d3-4474-4beb-9c7b-2d890d2134e6\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.368261 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.389975 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdnr4\" (UniqueName: \"kubernetes.io/projected/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-kube-api-access-mdnr4\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.390020 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.390089 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.390122 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-867fn\" (UniqueName: \"kubernetes.io/projected/d66b6939-c9a0-413b-b09a-5eb529657837-kube-api-access-867fn\") pod \"watcher-operator-controller-manager-667bd8d554-x98sw\" (UID: \"d66b6939-c9a0-413b-b09a-5eb529657837\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.426870 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.428225 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.433568 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-b2cv9" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.447832 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.495679 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.495734 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-867fn\" (UniqueName: \"kubernetes.io/projected/d66b6939-c9a0-413b-b09a-5eb529657837-kube-api-access-867fn\") pod \"watcher-operator-controller-manager-667bd8d554-x98sw\" (UID: \"d66b6939-c9a0-413b-b09a-5eb529657837\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.496762 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77knq\" (UniqueName: \"kubernetes.io/projected/3dccc54c-41e2-4745-839c-df51ca4e1825-kube-api-access-77knq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-69rd4\" (UID: \"3dccc54c-41e2-4745-839c-df51ca4e1825\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4" Dec 10 09:48:21 crc kubenswrapper[4715]: E1210 09:48:21.497008 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 09:48:21 crc kubenswrapper[4715]: E1210 09:48:21.497081 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:21.99706238 +0000 UTC m=+864.740608631 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "metrics-server-cert" not found Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.497204 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdnr4\" (UniqueName: \"kubernetes.io/projected/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-kube-api-access-mdnr4\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.497354 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:21 crc kubenswrapper[4715]: E1210 09:48:21.498071 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 09:48:21 crc kubenswrapper[4715]: E1210 09:48:21.498181 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:21.99813549 +0000 UTC m=+864.741681741 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "webhook-server-cert" not found Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.498247 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.507013 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.521380 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdnr4\" (UniqueName: \"kubernetes.io/projected/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-kube-api-access-mdnr4\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.539849 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.599019 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77knq\" (UniqueName: \"kubernetes.io/projected/3dccc54c-41e2-4745-839c-df51ca4e1825-kube-api-access-77knq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-69rd4\" (UID: \"3dccc54c-41e2-4745-839c-df51ca4e1825\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.599098 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:21 crc kubenswrapper[4715]: E1210 09:48:21.599284 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 09:48:21 crc kubenswrapper[4715]: E1210 09:48:21.599327 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert podName:d329770a-7d39-41d0-b4b1-a52e2e9ccda7 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:22.599313472 +0000 UTC m=+865.342859723 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fphth6" (UID: "d329770a-7d39-41d0-b4b1-a52e2e9ccda7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.608405 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-867fn\" (UniqueName: \"kubernetes.io/projected/d66b6939-c9a0-413b-b09a-5eb529657837-kube-api-access-867fn\") pod \"watcher-operator-controller-manager-667bd8d554-x98sw\" (UID: \"d66b6939-c9a0-413b-b09a-5eb529657837\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.639434 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77knq\" (UniqueName: \"kubernetes.io/projected/3dccc54c-41e2-4745-839c-df51ca4e1825-kube-api-access-77knq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-69rd4\" (UID: \"3dccc54c-41e2-4745-839c-df51ca4e1825\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.639716 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.689682 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.704618 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.834751 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4" Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.851296 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5"] Dec 10 09:48:21 crc kubenswrapper[4715]: I1210 09:48:21.948098 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg"] Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.058715 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.058818 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:22 crc kubenswrapper[4715]: E1210 09:48:22.058872 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 09:48:22 crc kubenswrapper[4715]: E1210 09:48:22.058975 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:23.058953921 +0000 UTC m=+865.802500172 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "webhook-server-cert" not found Dec 10 09:48:22 crc kubenswrapper[4715]: E1210 09:48:22.059002 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 09:48:22 crc kubenswrapper[4715]: E1210 09:48:22.059062 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:23.059043224 +0000 UTC m=+865.802589535 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "metrics-server-cert" not found Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.158488 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh"] Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.184849 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" event={"ID":"5949f2c9-f510-47f4-988c-cae06576536a","Type":"ContainerStarted","Data":"f215a0d9e5cbf1f0783ce8c0039356592dde767dc3b7da3705764dd3e75759d4"} Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.186551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" event={"ID":"f1e4cba4-0694-4477-b031-5caac1fbcb84","Type":"ContainerStarted","Data":"230989146b56dc242daa76041a5ca5cd4966e9bce6cf8c4f8c98af0cdb7b7597"} Dec 10 09:48:22 crc kubenswrapper[4715]: W1210 09:48:22.188280 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c14bbb7_a3cf_4654_ba2c_2d993022fa72.slice/crio-e928bcb61ffb99744742534aa176a4b951d9c0b805cdebec7b2094aec5456b6b WatchSource:0}: Error finding container e928bcb61ffb99744742534aa176a4b951d9c0b805cdebec7b2094aec5456b6b: Status 404 returned error can't find the container with id e928bcb61ffb99744742534aa176a4b951d9c0b805cdebec7b2094aec5456b6b Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.199106 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" event={"ID":"af548670-106c-4fbc-b6a5-102896191190","Type":"ContainerStarted","Data":"0d844838212163e8dff1271277eabea7526102aba62d0126685689036ef908e3"} Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.246282 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5"] Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.363073 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:22 crc kubenswrapper[4715]: E1210 09:48:22.363287 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 09:48:22 crc kubenswrapper[4715]: E1210 09:48:22.363384 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert podName:31619ce3-146d-4efb-898f-75bd1e007de9 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:24.363352423 +0000 UTC m=+867.106898674 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert") pod "infra-operator-controller-manager-78d48bff9d-mxxcb" (UID: "31619ce3-146d-4efb-898f-75bd1e007de9") : secret "infra-operator-webhook-server-cert" not found Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.667840 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:22 crc kubenswrapper[4715]: E1210 09:48:22.669458 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 09:48:22 crc kubenswrapper[4715]: E1210 09:48:22.669509 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert podName:d329770a-7d39-41d0-b4b1-a52e2e9ccda7 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:24.669491744 +0000 UTC m=+867.413037995 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fphth6" (UID: "d329770a-7d39-41d0-b4b1-a52e2e9ccda7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.748975 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn"] Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.753766 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q"] Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.759865 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt"] Dec 10 09:48:22 crc kubenswrapper[4715]: I1210 09:48:22.763529 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz"] Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.073348 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.073468 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.073519 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.073631 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:25.073605756 +0000 UTC m=+867.817152007 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "webhook-server-cert" not found Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.073673 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.073740 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:25.073719729 +0000 UTC m=+867.817265980 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "metrics-server-cert" not found Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.166540 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54"] Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.250082 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-m75xm"] Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.259738 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd"] Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.272469 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" event={"ID":"c396c828-63cf-4cd7-a050-c359ce8e1c8b","Type":"ContainerStarted","Data":"f7998ae68b8083fdb55f7660a20650f87a39126e344e7687924a239a166f4636"} Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.276749 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7"] Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.279492 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" event={"ID":"629d31d3-4474-4beb-9c7b-2d890d2134e6","Type":"ContainerStarted","Data":"3700a4297fd3e699845de2c16aea5e9946abe8fc29f929b72920893adf6c07fb"} Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.284852 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k"] Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.292100 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl"] Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.292407 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" event={"ID":"c3397611-a5cb-455c-8c5e-dd24ab61545a","Type":"ContainerStarted","Data":"02e77510ba5c2777cd8ca5c1e9a058db33a3fa6ee935541579887d5f510a9381"} Dec 10 09:48:23 crc kubenswrapper[4715]: W1210 09:48:23.303425 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfe88ca5_7d1d_4fff_8404_8b192ae48b09.slice/crio-136410b5ee9ce8e162e1484a6d6f587b61f26bcb61c722eaf2b65f600817676b WatchSource:0}: Error finding container 136410b5ee9ce8e162e1484a6d6f587b61f26bcb61c722eaf2b65f600817676b: Status 404 returned error can't find the container with id 136410b5ee9ce8e162e1484a6d6f587b61f26bcb61c722eaf2b65f600817676b Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.306819 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw"] Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.307319 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2z772,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-wcvvl_openstack-operators(cfe88ca5-7d1d-4fff-8404-8b192ae48b09): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.308635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" event={"ID":"20025d92-9314-436f-a05e-4bfee3d3add0","Type":"ContainerStarted","Data":"59de659018e0f3a01a292a2dc6d2f51fe55bbbbe8b8cd25b5faf251ed85bc8c3"} Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.309086 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2z772,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-wcvvl_openstack-operators(cfe88ca5-7d1d-4fff-8404-8b192ae48b09): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.310654 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" podUID="cfe88ca5-7d1d-4fff-8404-8b192ae48b09" Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.311611 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" event={"ID":"6c14bbb7-a3cf-4654-ba2c-2d993022fa72","Type":"ContainerStarted","Data":"e928bcb61ffb99744742534aa176a4b951d9c0b805cdebec7b2094aec5456b6b"} Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.312269 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lcj94,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-gbrf4_openstack-operators(91566e3d-fbed-450a-bc66-1bf34f98d57f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.312998 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" event={"ID":"75ebabe2-8b57-4eee-9b3b-b238fafd94b7","Type":"ContainerStarted","Data":"6fffca4fbc218775de8df557bd7156f661c03bbf4a09a50bc3ffff2323f54bb7"} Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.314230 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lcj94,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-gbrf4_openstack-operators(91566e3d-fbed-450a-bc66-1bf34f98d57f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.314695 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" event={"ID":"06228f2c-8f0c-493a-850f-b1e5fd41a23b","Type":"ContainerStarted","Data":"9a0b191e1819877b3ac683884d417e51ecbb01082dd2631c33642274d68f5e6a"} Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.315481 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" podUID="91566e3d-fbed-450a-bc66-1bf34f98d57f" Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.316702 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" event={"ID":"78ba49c9-f451-4982-8d88-6bf0444c8b60","Type":"ContainerStarted","Data":"70fe5a0f3dbd22b7769b9cfed8f360797e62829cf30e5df01238d09081961a35"} Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.322235 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc"] Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.329151 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rt6zc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-ht6n4_openstack-operators(67b3cab4-c74e-4861-875b-e38aac7ced71): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.331453 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rt6zc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-ht6n4_openstack-operators(67b3cab4-c74e-4861-875b-e38aac7ced71): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.331895 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4"] Dec 10 09:48:23 crc kubenswrapper[4715]: E1210 09:48:23.332663 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" podUID="67b3cab4-c74e-4861-875b-e38aac7ced71" Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.338405 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4"] Dec 10 09:48:23 crc kubenswrapper[4715]: I1210 09:48:23.344533 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4"] Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.331018 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" event={"ID":"91566e3d-fbed-450a-bc66-1bf34f98d57f","Type":"ContainerStarted","Data":"9659808acd192b10dadd527a6a8639fc0bc9a2ff9f9b8fd2710ae48e1fe1d8ff"} Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.334492 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" event={"ID":"d66b6939-c9a0-413b-b09a-5eb529657837","Type":"ContainerStarted","Data":"82c14c217cc25364bbea351afde5a911bb562880ca7472c03591f73e1bd2138a"} Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.336057 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" event={"ID":"63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd","Type":"ContainerStarted","Data":"804d69b955e475fc71533889eec27235907cfe457014dcdb2c79e648e0aecbe0"} Dec 10 09:48:24 crc kubenswrapper[4715]: E1210 09:48:24.336613 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" podUID="91566e3d-fbed-450a-bc66-1bf34f98d57f" Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.337981 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" event={"ID":"d80992c0-9c12-4266-b705-6eaed7e46d1e","Type":"ContainerStarted","Data":"161ceae8471b4d65958eb27f7532ade24fe12b6b632657e4dd417e72d1761a37"} Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.341381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" event={"ID":"cfe88ca5-7d1d-4fff-8404-8b192ae48b09","Type":"ContainerStarted","Data":"136410b5ee9ce8e162e1484a6d6f587b61f26bcb61c722eaf2b65f600817676b"} Dec 10 09:48:24 crc kubenswrapper[4715]: E1210 09:48:24.346138 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" podUID="cfe88ca5-7d1d-4fff-8404-8b192ae48b09" Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.349248 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" event={"ID":"50f2991b-7b13-4d84-8a4d-504d9b873023","Type":"ContainerStarted","Data":"74ccf1856cae3406dd11b61f1731e587a5e6be1a11d25063abf95519476680ec"} Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.352352 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" event={"ID":"67b3cab4-c74e-4861-875b-e38aac7ced71","Type":"ContainerStarted","Data":"b8b191d7707f32eadf765aa00e89dfe779cc55efc1790efcde74931cd01d8705"} Dec 10 09:48:24 crc kubenswrapper[4715]: E1210 09:48:24.368437 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" podUID="67b3cab4-c74e-4861-875b-e38aac7ced71" Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.369715 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" event={"ID":"01645280-6af5-4f8d-b412-3d7e0efa9054","Type":"ContainerStarted","Data":"05e206212dd054fb61d64829166e9b1a079355af8b65fb17b9965865b2f75991"} Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.372806 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4" event={"ID":"3dccc54c-41e2-4745-839c-df51ca4e1825","Type":"ContainerStarted","Data":"9746fb9dc7f46332a4779e167e4654acd4174d552db7b79e1ac51bc960bf1b4b"} Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.398447 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:24 crc kubenswrapper[4715]: E1210 09:48:24.399133 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 09:48:24 crc kubenswrapper[4715]: E1210 09:48:24.399217 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert podName:31619ce3-146d-4efb-898f-75bd1e007de9 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:28.399193178 +0000 UTC m=+871.142739429 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert") pod "infra-operator-controller-manager-78d48bff9d-mxxcb" (UID: "31619ce3-146d-4efb-898f-75bd1e007de9") : secret "infra-operator-webhook-server-cert" not found Dec 10 09:48:24 crc kubenswrapper[4715]: I1210 09:48:24.705960 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:24 crc kubenswrapper[4715]: E1210 09:48:24.706691 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 09:48:24 crc kubenswrapper[4715]: E1210 09:48:24.706770 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert podName:d329770a-7d39-41d0-b4b1-a52e2e9ccda7 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:28.706750559 +0000 UTC m=+871.450296810 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fphth6" (UID: "d329770a-7d39-41d0-b4b1-a52e2e9ccda7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 09:48:25 crc kubenswrapper[4715]: I1210 09:48:25.157650 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:25 crc kubenswrapper[4715]: I1210 09:48:25.157988 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:25 crc kubenswrapper[4715]: E1210 09:48:25.158375 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 09:48:25 crc kubenswrapper[4715]: E1210 09:48:25.160124 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:29.160099779 +0000 UTC m=+871.903646030 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "metrics-server-cert" not found Dec 10 09:48:25 crc kubenswrapper[4715]: E1210 09:48:25.158436 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 09:48:25 crc kubenswrapper[4715]: E1210 09:48:25.161330 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:29.161296623 +0000 UTC m=+871.904842944 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "webhook-server-cert" not found Dec 10 09:48:25 crc kubenswrapper[4715]: E1210 09:48:25.405223 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" podUID="91566e3d-fbed-450a-bc66-1bf34f98d57f" Dec 10 09:48:25 crc kubenswrapper[4715]: E1210 09:48:25.405645 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" podUID="67b3cab4-c74e-4861-875b-e38aac7ced71" Dec 10 09:48:25 crc kubenswrapper[4715]: E1210 09:48:25.406081 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" podUID="cfe88ca5-7d1d-4fff-8404-8b192ae48b09" Dec 10 09:48:28 crc kubenswrapper[4715]: I1210 09:48:28.464382 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:28 crc kubenswrapper[4715]: E1210 09:48:28.464867 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 09:48:28 crc kubenswrapper[4715]: E1210 09:48:28.464939 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert podName:31619ce3-146d-4efb-898f-75bd1e007de9 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:36.464903488 +0000 UTC m=+879.208449739 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert") pod "infra-operator-controller-manager-78d48bff9d-mxxcb" (UID: "31619ce3-146d-4efb-898f-75bd1e007de9") : secret "infra-operator-webhook-server-cert" not found Dec 10 09:48:28 crc kubenswrapper[4715]: I1210 09:48:28.779013 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:28 crc kubenswrapper[4715]: E1210 09:48:28.779183 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 09:48:28 crc kubenswrapper[4715]: E1210 09:48:28.779254 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert podName:d329770a-7d39-41d0-b4b1-a52e2e9ccda7 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:36.779235542 +0000 UTC m=+879.522781793 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fphth6" (UID: "d329770a-7d39-41d0-b4b1-a52e2e9ccda7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 09:48:29 crc kubenswrapper[4715]: I1210 09:48:29.185089 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:29 crc kubenswrapper[4715]: I1210 09:48:29.185160 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:29 crc kubenswrapper[4715]: E1210 09:48:29.185294 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 09:48:29 crc kubenswrapper[4715]: E1210 09:48:29.185341 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:37.185325631 +0000 UTC m=+879.928871882 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "metrics-server-cert" not found Dec 10 09:48:29 crc kubenswrapper[4715]: E1210 09:48:29.185508 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 09:48:29 crc kubenswrapper[4715]: E1210 09:48:29.185582 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:37.185563088 +0000 UTC m=+879.929109339 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "webhook-server-cert" not found Dec 10 09:48:36 crc kubenswrapper[4715]: E1210 09:48:36.068776 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87" Dec 10 09:48:36 crc kubenswrapper[4715]: E1210 09:48:36.070290 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m67tk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-h9rpn_openstack-operators(c3397611-a5cb-455c-8c5e-dd24ab61545a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:48:36 crc kubenswrapper[4715]: I1210 09:48:36.536576 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:36 crc kubenswrapper[4715]: I1210 09:48:36.555199 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31619ce3-146d-4efb-898f-75bd1e007de9-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxxcb\" (UID: \"31619ce3-146d-4efb-898f-75bd1e007de9\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:36 crc kubenswrapper[4715]: E1210 09:48:36.739654 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 10 09:48:36 crc kubenswrapper[4715]: E1210 09:48:36.740015 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m9288,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-d875q_openstack-operators(06228f2c-8f0c-493a-850f-b1e5fd41a23b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:48:36 crc kubenswrapper[4715]: I1210 09:48:36.755510 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:48:36 crc kubenswrapper[4715]: I1210 09:48:36.840864 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:36 crc kubenswrapper[4715]: I1210 09:48:36.850907 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d329770a-7d39-41d0-b4b1-a52e2e9ccda7-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fphth6\" (UID: \"d329770a-7d39-41d0-b4b1-a52e2e9ccda7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:36 crc kubenswrapper[4715]: I1210 09:48:36.978421 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:48:37 crc kubenswrapper[4715]: I1210 09:48:37.246181 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:37 crc kubenswrapper[4715]: I1210 09:48:37.246254 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:37 crc kubenswrapper[4715]: E1210 09:48:37.246860 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 09:48:37 crc kubenswrapper[4715]: E1210 09:48:37.246941 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs podName:8c747bc0-5cc3-4cd2-8283-bdb2f2742167 nodeName:}" failed. No retries permitted until 2025-12-10 09:48:53.246924681 +0000 UTC m=+895.990470932 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs") pod "openstack-operator-controller-manager-7fd7979c46-c59rw" (UID: "8c747bc0-5cc3-4cd2-8283-bdb2f2742167") : secret "webhook-server-cert" not found Dec 10 09:48:37 crc kubenswrapper[4715]: I1210 09:48:37.252869 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-metrics-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:37 crc kubenswrapper[4715]: E1210 09:48:37.274532 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 10 09:48:37 crc kubenswrapper[4715]: E1210 09:48:37.275015 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m2k75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-88z54_openstack-operators(75ebabe2-8b57-4eee-9b3b-b238fafd94b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:48:39 crc kubenswrapper[4715]: E1210 09:48:39.803950 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 10 09:48:39 crc kubenswrapper[4715]: E1210 09:48:39.804458 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zvj66,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-scbxh_openstack-operators(6c14bbb7-a3cf-4654-ba2c-2d993022fa72): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:48:42 crc kubenswrapper[4715]: E1210 09:48:42.281403 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad" Dec 10 09:48:42 crc kubenswrapper[4715]: E1210 09:48:42.281591 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wxjlq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-tjqbt_openstack-operators(78ba49c9-f451-4982-8d88-6bf0444c8b60): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:48:43 crc kubenswrapper[4715]: E1210 09:48:43.374565 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 10 09:48:43 crc kubenswrapper[4715]: E1210 09:48:43.375114 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8b5qg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-66fzd_openstack-operators(01645280-6af5-4f8d-b412-3d7e0efa9054): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:48:44 crc kubenswrapper[4715]: E1210 09:48:44.560843 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f" Dec 10 09:48:44 crc kubenswrapper[4715]: E1210 09:48:44.561082 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-djjjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-2957k_openstack-operators(63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:48:46 crc kubenswrapper[4715]: E1210 09:48:46.593429 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a" Dec 10 09:48:46 crc kubenswrapper[4715]: E1210 09:48:46.593985 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-swfpw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-p69tc_openstack-operators(d80992c0-9c12-4266-b705-6eaed7e46d1e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:48:53 crc kubenswrapper[4715]: I1210 09:48:53.266548 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:53 crc kubenswrapper[4715]: I1210 09:48:53.274419 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8c747bc0-5cc3-4cd2-8283-bdb2f2742167-webhook-certs\") pod \"openstack-operator-controller-manager-7fd7979c46-c59rw\" (UID: \"8c747bc0-5cc3-4cd2-8283-bdb2f2742167\") " pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:48:53 crc kubenswrapper[4715]: I1210 09:48:53.511282 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:49:02 crc kubenswrapper[4715]: E1210 09:49:02.307758 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 10 09:49:02 crc kubenswrapper[4715]: E1210 09:49:02.308234 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2z772,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-wcvvl_openstack-operators(cfe88ca5-7d1d-4fff-8404-8b192ae48b09): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.050200 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.050419 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rt6zc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-ht6n4_openstack-operators(67b3cab4-c74e-4861-875b-e38aac7ced71): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.523781 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.524021 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-77knq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-69rd4_openstack-operators(3dccc54c-41e2-4745-839c-df51ca4e1825): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.525649 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4" podUID="3dccc54c-41e2-4745-839c-df51ca4e1825" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.570430 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.570627 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-djjjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-2957k_openstack-operators(63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.571827 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" podUID="63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.705431 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.708792 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8b5qg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-66fzd_openstack-operators(01645280-6af5-4f8d-b412-3d7e0efa9054): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.710317 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" podUID="01645280-6af5-4f8d-b412-3d7e0efa9054" Dec 10 09:49:03 crc kubenswrapper[4715]: E1210 09:49:03.715530 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4" podUID="3dccc54c-41e2-4745-839c-df51ca4e1825" Dec 10 09:49:03 crc kubenswrapper[4715]: I1210 09:49:03.987189 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb"] Dec 10 09:49:04 crc kubenswrapper[4715]: W1210 09:49:04.041775 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31619ce3_146d_4efb_898f_75bd1e007de9.slice/crio-cf86610134ab7c663cc32cd248490c955b1137673d56450bb400bfefd117d1c0 WatchSource:0}: Error finding container cf86610134ab7c663cc32cd248490c955b1137673d56450bb400bfefd117d1c0: Status 404 returned error can't find the container with id cf86610134ab7c663cc32cd248490c955b1137673d56450bb400bfefd117d1c0 Dec 10 09:49:04 crc kubenswrapper[4715]: I1210 09:49:04.098969 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6"] Dec 10 09:49:04 crc kubenswrapper[4715]: W1210 09:49:04.351298 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd329770a_7d39_41d0_b4b1_a52e2e9ccda7.slice/crio-b5bd507557f3be0ebd2af0bcff4060a79a54b9adc058ebdce88385afc8a19d22 WatchSource:0}: Error finding container b5bd507557f3be0ebd2af0bcff4060a79a54b9adc058ebdce88385afc8a19d22: Status 404 returned error can't find the container with id b5bd507557f3be0ebd2af0bcff4060a79a54b9adc058ebdce88385afc8a19d22 Dec 10 09:49:04 crc kubenswrapper[4715]: E1210 09:49:04.371885 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 09:49:04 crc kubenswrapper[4715]: E1210 09:49:04.372069 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-swfpw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-p69tc_openstack-operators(d80992c0-9c12-4266-b705-6eaed7e46d1e): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 09:49:04 crc kubenswrapper[4715]: E1210 09:49:04.373795 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" podUID="d80992c0-9c12-4266-b705-6eaed7e46d1e" Dec 10 09:49:04 crc kubenswrapper[4715]: E1210 09:49:04.381925 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 09:49:04 crc kubenswrapper[4715]: E1210 09:49:04.382100 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wxjlq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-tjqbt_openstack-operators(78ba49c9-f451-4982-8d88-6bf0444c8b60): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 09:49:04 crc kubenswrapper[4715]: E1210 09:49:04.383390 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" podUID="78ba49c9-f451-4982-8d88-6bf0444c8b60" Dec 10 09:49:04 crc kubenswrapper[4715]: I1210 09:49:04.573720 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw"] Dec 10 09:49:04 crc kubenswrapper[4715]: I1210 09:49:04.719185 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" event={"ID":"d329770a-7d39-41d0-b4b1-a52e2e9ccda7","Type":"ContainerStarted","Data":"b5bd507557f3be0ebd2af0bcff4060a79a54b9adc058ebdce88385afc8a19d22"} Dec 10 09:49:04 crc kubenswrapper[4715]: I1210 09:49:04.721167 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" event={"ID":"629d31d3-4474-4beb-9c7b-2d890d2134e6","Type":"ContainerStarted","Data":"54f3f4b7705f82e419c23d0fc13192e6b1f492f28b8842faa0c37943fe8f36dd"} Dec 10 09:49:04 crc kubenswrapper[4715]: I1210 09:49:04.722789 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" event={"ID":"af548670-106c-4fbc-b6a5-102896191190","Type":"ContainerStarted","Data":"f68b13598cc0166358a029ac710671c55540e06ce78009c3287ef77199debe5d"} Dec 10 09:49:04 crc kubenswrapper[4715]: I1210 09:49:04.724098 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" event={"ID":"8c747bc0-5cc3-4cd2-8283-bdb2f2742167","Type":"ContainerStarted","Data":"4d78fbee39cf0f250e2356116585624760242d4c7692b5b2adda856c8f713a5c"} Dec 10 09:49:04 crc kubenswrapper[4715]: I1210 09:49:04.725345 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" event={"ID":"31619ce3-146d-4efb-898f-75bd1e007de9","Type":"ContainerStarted","Data":"cf86610134ab7c663cc32cd248490c955b1137673d56450bb400bfefd117d1c0"} Dec 10 09:49:04 crc kubenswrapper[4715]: I1210 09:49:04.728851 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" event={"ID":"f1e4cba4-0694-4477-b031-5caac1fbcb84","Type":"ContainerStarted","Data":"9051435971b360dd0a62a0cc73b0f99266c2654e520d34a681a9a73b6ec080f7"} Dec 10 09:49:05 crc kubenswrapper[4715]: E1210 09:49:05.095507 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 09:49:05 crc kubenswrapper[4715]: E1210 09:49:05.095700 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zvj66,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-scbxh_openstack-operators(6c14bbb7-a3cf-4654-ba2c-2d993022fa72): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 09:49:05 crc kubenswrapper[4715]: E1210 09:49:05.096869 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" podUID="6c14bbb7-a3cf-4654-ba2c-2d993022fa72" Dec 10 09:49:05 crc kubenswrapper[4715]: E1210 09:49:05.401561 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 09:49:05 crc kubenswrapper[4715]: E1210 09:49:05.402184 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m67tk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-h9rpn_openstack-operators(c3397611-a5cb-455c-8c5e-dd24ab61545a): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" logger="UnhandledError" Dec 10 09:49:05 crc kubenswrapper[4715]: E1210 09:49:05.403372 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\\\": context canceled\"]" pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" podUID="c3397611-a5cb-455c-8c5e-dd24ab61545a" Dec 10 09:49:05 crc kubenswrapper[4715]: E1210 09:49:05.711208 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 09:49:05 crc kubenswrapper[4715]: E1210 09:49:05.711349 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m9288,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-d875q_openstack-operators(06228f2c-8f0c-493a-850f-b1e5fd41a23b): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce\": context canceled" logger="UnhandledError" Dec 10 09:49:05 crc kubenswrapper[4715]: E1210 09:49:05.713300 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce\\\": context canceled\"]" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" podUID="06228f2c-8f0c-493a-850f-b1e5fd41a23b" Dec 10 09:49:05 crc kubenswrapper[4715]: I1210 09:49:05.740055 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" event={"ID":"c396c828-63cf-4cd7-a050-c359ce8e1c8b","Type":"ContainerStarted","Data":"996a2e412fe55f745a370ebd70af6fba7c48c6f2b8b418fdd3c0c16d32b054e4"} Dec 10 09:49:05 crc kubenswrapper[4715]: I1210 09:49:05.748448 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" event={"ID":"5949f2c9-f510-47f4-988c-cae06576536a","Type":"ContainerStarted","Data":"ab0f04635407bbc1be7680dff953831e89774fbd03ae524876e9029e58322b8b"} Dec 10 09:49:05 crc kubenswrapper[4715]: I1210 09:49:05.750800 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" event={"ID":"50f2991b-7b13-4d84-8a4d-504d9b873023","Type":"ContainerStarted","Data":"a5efbd46d72df6c43d281cdb9d67e29592a20d27e9d65a3e17d079f3e3a1cddd"} Dec 10 09:49:05 crc kubenswrapper[4715]: I1210 09:49:05.752650 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" event={"ID":"20025d92-9314-436f-a05e-4bfee3d3add0","Type":"ContainerStarted","Data":"e8aaaf68773a4272bd88718b474852b81941a6ed4b4847522aeb42cf1fa975df"} Dec 10 09:49:05 crc kubenswrapper[4715]: I1210 09:49:05.757887 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" event={"ID":"d66b6939-c9a0-413b-b09a-5eb529657837","Type":"ContainerStarted","Data":"1f5d0b5f6a6fe501adf1366c0013a478a0834a3a1495d0a23b53ae947d9868cc"} Dec 10 09:49:09 crc kubenswrapper[4715]: I1210 09:49:09.955802 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" event={"ID":"91566e3d-fbed-450a-bc66-1bf34f98d57f","Type":"ContainerStarted","Data":"146f0a0c767695fdba20a33977e5d571f39b7e945dbeaf944b1a39a45c2b1953"} Dec 10 09:49:10 crc kubenswrapper[4715]: I1210 09:49:09.957970 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" event={"ID":"8c747bc0-5cc3-4cd2-8283-bdb2f2742167","Type":"ContainerStarted","Data":"9ce6e36e3cbef12a905e13b0238c5208a537c297d4cd138f48174da3c1643c74"} Dec 10 09:49:10 crc kubenswrapper[4715]: I1210 09:49:09.959068 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:49:10 crc kubenswrapper[4715]: I1210 09:49:10.181489 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" podStartSLOduration=49.18147246 podStartE2EDuration="49.18147246s" podCreationTimestamp="2025-12-10 09:48:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:49:10.17557546 +0000 UTC m=+912.919121701" watchObservedRunningTime="2025-12-10 09:49:10.18147246 +0000 UTC m=+912.925018711" Dec 10 09:49:10 crc kubenswrapper[4715]: E1210 09:49:10.622538 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" podUID="cfe88ca5-7d1d-4fff-8404-8b192ae48b09" Dec 10 09:49:11 crc kubenswrapper[4715]: E1210 09:49:11.023948 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" podUID="67b3cab4-c74e-4861-875b-e38aac7ced71" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.024411 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" event={"ID":"af548670-106c-4fbc-b6a5-102896191190","Type":"ContainerStarted","Data":"1e904943f70fa082d460d9c02d245d2829305a54bf12922ec136a7dc06f8f326"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.024689 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.062857 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.082733 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" event={"ID":"91566e3d-fbed-450a-bc66-1bf34f98d57f","Type":"ContainerStarted","Data":"6cfbf5db2ba221a256e3b4278c1384c67f8769a16fdf92d0be24499b1f1acd8e"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.082843 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.156202 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-htklg" podStartSLOduration=3.266498346 podStartE2EDuration="51.156172041s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:22.006022929 +0000 UTC m=+864.749569180" lastFinishedPulling="2025-12-10 09:49:09.895696614 +0000 UTC m=+912.639242875" observedRunningTime="2025-12-10 09:49:11.126105557 +0000 UTC m=+913.869651808" watchObservedRunningTime="2025-12-10 09:49:11.156172041 +0000 UTC m=+913.899718292" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.292374 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" event={"ID":"78ba49c9-f451-4982-8d88-6bf0444c8b60","Type":"ContainerStarted","Data":"0edb321b148979830f8fd8a65f21c7fd92debf70420f912ea5df6a7bc87e9ad4"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.314407 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" event={"ID":"c3397611-a5cb-455c-8c5e-dd24ab61545a","Type":"ContainerStarted","Data":"0908abe323adcedb357f0c129318060fffb4e7e9dc20dabb06d31af4204175eb"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.327976 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" podStartSLOduration=10.636411506 podStartE2EDuration="51.327889411s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.312179409 +0000 UTC m=+866.055725660" lastFinishedPulling="2025-12-10 09:49:04.003657304 +0000 UTC m=+906.747203565" observedRunningTime="2025-12-10 09:49:11.291066041 +0000 UTC m=+914.034612302" watchObservedRunningTime="2025-12-10 09:49:11.327889411 +0000 UTC m=+914.071435692" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.349095 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" event={"ID":"cfe88ca5-7d1d-4fff-8404-8b192ae48b09","Type":"ContainerStarted","Data":"271ee8a6132fcd77ea85d61bcbc11654613e40042a27f926539ebc75fbe1ef7c"} Dec 10 09:49:11 crc kubenswrapper[4715]: E1210 09:49:11.357889 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" podUID="cfe88ca5-7d1d-4fff-8404-8b192ae48b09" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.363309 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" event={"ID":"31619ce3-146d-4efb-898f-75bd1e007de9","Type":"ContainerStarted","Data":"3b5ce2b8db59d29271eb9932bd38bdea6a3fb839f3cda2ce2ac29d58f55b35c8"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.398402 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" event={"ID":"63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd","Type":"ContainerStarted","Data":"2b1a67cb0b3d54fd58c448bab9ec7face9cf3d5eef383ae4b1076750066461ad"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.423455 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" event={"ID":"d329770a-7d39-41d0-b4b1-a52e2e9ccda7","Type":"ContainerStarted","Data":"130f148a28729e758b9fafa6d9f4ce6068c7bbec88927e95f6e3447dcf9f1233"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.429943 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" event={"ID":"5949f2c9-f510-47f4-988c-cae06576536a","Type":"ContainerStarted","Data":"058beb6e3f2a2d4af6e3ba901b39aca087879aeb6a0f84e3125047b6529b6070"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.432720 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.436530 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.437692 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" event={"ID":"06228f2c-8f0c-493a-850f-b1e5fd41a23b","Type":"ContainerStarted","Data":"f04ab0a3264b048046540f78bcfceb441d92d5a3088b73b10167ef7c9de9d4bc"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.440248 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" event={"ID":"20025d92-9314-436f-a05e-4bfee3d3add0","Type":"ContainerStarted","Data":"d8174537716a897ca9d356d5fdb576afdf05e8ded17633a5c3e77b1abc5fe0a3"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.442042 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.443091 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.503069 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" event={"ID":"c396c828-63cf-4cd7-a050-c359ce8e1c8b","Type":"ContainerStarted","Data":"50c54509b664ac55b98c99573f40a82b50a3f5975e2387b813963c0a22245835"} Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.504080 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.510223 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.511424 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.515312 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-j2drs" podStartSLOduration=3.406052104 podStartE2EDuration="51.51529092s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:21.8620018 +0000 UTC m=+864.605548051" lastFinishedPulling="2025-12-10 09:49:09.971240616 +0000 UTC m=+912.714786867" observedRunningTime="2025-12-10 09:49:11.506684304 +0000 UTC m=+914.250230555" watchObservedRunningTime="2025-12-10 09:49:11.51529092 +0000 UTC m=+914.258837171" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.527770 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.529501 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-z25gz" podStartSLOduration=4.41391861 podStartE2EDuration="51.529482399s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:22.780199247 +0000 UTC m=+865.523745498" lastFinishedPulling="2025-12-10 09:49:09.895763046 +0000 UTC m=+912.639309287" observedRunningTime="2025-12-10 09:49:11.526412075 +0000 UTC m=+914.269958336" watchObservedRunningTime="2025-12-10 09:49:11.529482399 +0000 UTC m=+914.273028650" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.594255 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jdcr5" podStartSLOduration=4.048988878 podStartE2EDuration="51.594236275s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:22.310003658 +0000 UTC m=+865.053549909" lastFinishedPulling="2025-12-10 09:49:09.855251055 +0000 UTC m=+912.598797306" observedRunningTime="2025-12-10 09:49:11.550930697 +0000 UTC m=+914.294476958" watchObservedRunningTime="2025-12-10 09:49:11.594236275 +0000 UTC m=+914.337782516" Dec 10 09:49:11 crc kubenswrapper[4715]: I1210 09:49:11.677597 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" podStartSLOduration=4.964481848 podStartE2EDuration="51.677577701s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.182754865 +0000 UTC m=+865.926301116" lastFinishedPulling="2025-12-10 09:49:09.895850718 +0000 UTC m=+912.639396969" observedRunningTime="2025-12-10 09:49:11.675302018 +0000 UTC m=+914.418848269" watchObservedRunningTime="2025-12-10 09:49:11.677577701 +0000 UTC m=+914.421123972" Dec 10 09:49:11 crc kubenswrapper[4715]: E1210 09:49:11.749205 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" podUID="75ebabe2-8b57-4eee-9b3b-b238fafd94b7" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.519926 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" event={"ID":"d329770a-7d39-41d0-b4b1-a52e2e9ccda7","Type":"ContainerStarted","Data":"7ffcf38141b0e1141c053c4f5a4dea8e7fd1ffb7c4a38ee6c994b88a7bba84cf"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.521222 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.524291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" event={"ID":"c3397611-a5cb-455c-8c5e-dd24ab61545a","Type":"ContainerStarted","Data":"5fb984f363e9046469d26b81b50521c52504976cc9807c999f4bf4a2346c0944"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.524365 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.526866 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" event={"ID":"06228f2c-8f0c-493a-850f-b1e5fd41a23b","Type":"ContainerStarted","Data":"d40cae8c88a3f84d8cd774b2baf5176fcfac49b90f0312cc2f0c5db24fae2b2c"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.527057 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.528998 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" event={"ID":"50f2991b-7b13-4d84-8a4d-504d9b873023","Type":"ContainerStarted","Data":"214e54acb203375527fb3396f0ce351cf3d6d2a250777977adf2b74b07e7417f"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.529485 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.530838 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" event={"ID":"78ba49c9-f451-4982-8d88-6bf0444c8b60","Type":"ContainerStarted","Data":"680cd357807607531700bebeec663ec4ca90f7bc56c0e97230123b743cfb798b"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.531034 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.532616 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" event={"ID":"31619ce3-146d-4efb-898f-75bd1e007de9","Type":"ContainerStarted","Data":"d139fe4cefe2415a8004488a4eaf88a5b8f708f983688da02a043b168a165312"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.533274 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.534878 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" event={"ID":"01645280-6af5-4f8d-b412-3d7e0efa9054","Type":"ContainerStarted","Data":"97e731b53d071cefb5a0c3582a40279eb9c837671c23076d5c3b9a7d1f9c8f04"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.534897 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" event={"ID":"01645280-6af5-4f8d-b412-3d7e0efa9054","Type":"ContainerStarted","Data":"23ebbc316b623e2ea88d3c163f096340e2e66906258571457709c407c40b61a3"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.535291 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.540764 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" event={"ID":"6c14bbb7-a3cf-4654-ba2c-2d993022fa72","Type":"ContainerStarted","Data":"2590a4ac23ebe17b172864d8ae252709d7535513bc613ea34ac604d188500f07"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.540792 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" event={"ID":"6c14bbb7-a3cf-4654-ba2c-2d993022fa72","Type":"ContainerStarted","Data":"e35d7828d5b23870c320e9a7b2e20e77f5c9368317a14911233feff371e5b90f"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.541353 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.542945 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" event={"ID":"75ebabe2-8b57-4eee-9b3b-b238fafd94b7","Type":"ContainerStarted","Data":"9f3b53da80bd12213d3844ca79d1244ed2a7a2de027a4328ef1e7fc04ff09047"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.546361 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" event={"ID":"63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd","Type":"ContainerStarted","Data":"35c22dbf443cae43270b59123421330a8e25d9ee706ece4c9e734742ca4d2e73"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.546737 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.548079 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" event={"ID":"d80992c0-9c12-4266-b705-6eaed7e46d1e","Type":"ContainerStarted","Data":"8ab64db8dc2b637aa21df3fbee849daf5415c30655552b27b06a761392e420ce"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.548104 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" event={"ID":"d80992c0-9c12-4266-b705-6eaed7e46d1e","Type":"ContainerStarted","Data":"e6b64906fe1372f4f9df207faf7b634ce27a04045fd03a310344f836a59b83ef"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.548502 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.550661 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" event={"ID":"67b3cab4-c74e-4861-875b-e38aac7ced71","Type":"ContainerStarted","Data":"f714db12e672bcd747e0773dcfcf53ae58ef8ee2fc2d02438fae12b8469c469b"} Dec 10 09:49:12 crc kubenswrapper[4715]: E1210 09:49:12.552896 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" podUID="67b3cab4-c74e-4861-875b-e38aac7ced71" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.553066 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.554399 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" event={"ID":"f1e4cba4-0694-4477-b031-5caac1fbcb84","Type":"ContainerStarted","Data":"c1092ace9dda76e4017299395cd29ecd60e6ac3ad7e00edd0a4d4dab84a206d8"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.554625 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.555996 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" event={"ID":"d66b6939-c9a0-413b-b09a-5eb529657837","Type":"ContainerStarted","Data":"80c060b7ccf5d1246ab023ae7ed41840d49ba043cd9a8b52a7122d06c6f2e4cc"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.556815 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.557899 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.560527 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-m75xm" event={"ID":"629d31d3-4474-4beb-9c7b-2d890d2134e6","Type":"ContainerStarted","Data":"665f5d699b455d09953676ff254fd1063107d62dd80cff9702765311b7a1e641"} Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.676544 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.738004 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" podStartSLOduration=47.37819851 podStartE2EDuration="52.737983802s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:49:04.356166812 +0000 UTC m=+907.099713103" lastFinishedPulling="2025-12-10 09:49:09.715952144 +0000 UTC m=+912.459498395" observedRunningTime="2025-12-10 09:49:12.702291253 +0000 UTC m=+915.445837504" watchObservedRunningTime="2025-12-10 09:49:12.737983802 +0000 UTC m=+915.481530053" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.738469 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-jwwm5" podStartSLOduration=4.620987023 podStartE2EDuration="52.738461235s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:21.861551517 +0000 UTC m=+864.605097768" lastFinishedPulling="2025-12-10 09:49:09.979025729 +0000 UTC m=+912.722571980" observedRunningTime="2025-12-10 09:49:12.728737769 +0000 UTC m=+915.472284020" watchObservedRunningTime="2025-12-10 09:49:12.738461235 +0000 UTC m=+915.482007486" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.803403 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" podStartSLOduration=47.185995222 podStartE2EDuration="52.803386156s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:49:04.04930536 +0000 UTC m=+906.792851611" lastFinishedPulling="2025-12-10 09:49:09.666696284 +0000 UTC m=+912.410242545" observedRunningTime="2025-12-10 09:49:12.788011994 +0000 UTC m=+915.531558245" watchObservedRunningTime="2025-12-10 09:49:12.803386156 +0000 UTC m=+915.546932407" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.830042 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" podStartSLOduration=6.447561703 podStartE2EDuration="52.830025937s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.299086037 +0000 UTC m=+866.042632288" lastFinishedPulling="2025-12-10 09:49:09.681550271 +0000 UTC m=+912.425096522" observedRunningTime="2025-12-10 09:49:12.819158578 +0000 UTC m=+915.562704839" watchObservedRunningTime="2025-12-10 09:49:12.830025937 +0000 UTC m=+915.573572178" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.883294 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" podStartSLOduration=5.108641636 podStartE2EDuration="52.883275887s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:22.202743773 +0000 UTC m=+864.946290024" lastFinishedPulling="2025-12-10 09:49:09.977378024 +0000 UTC m=+912.720924275" observedRunningTime="2025-12-10 09:49:12.873203061 +0000 UTC m=+915.616749312" watchObservedRunningTime="2025-12-10 09:49:12.883275887 +0000 UTC m=+915.626822128" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.938237 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" podStartSLOduration=11.142501443 podStartE2EDuration="52.938217254s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.301835475 +0000 UTC m=+866.045381726" lastFinishedPulling="2025-12-10 09:49:05.097551286 +0000 UTC m=+907.841097537" observedRunningTime="2025-12-10 09:49:12.933304459 +0000 UTC m=+915.676850710" watchObservedRunningTime="2025-12-10 09:49:12.938217254 +0000 UTC m=+915.681763505" Dec 10 09:49:12 crc kubenswrapper[4715]: I1210 09:49:12.964504 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" podStartSLOduration=5.818658815 podStartE2EDuration="52.964488414s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:22.780150265 +0000 UTC m=+865.523696516" lastFinishedPulling="2025-12-10 09:49:09.925979864 +0000 UTC m=+912.669526115" observedRunningTime="2025-12-10 09:49:12.963523098 +0000 UTC m=+915.707069349" watchObservedRunningTime="2025-12-10 09:49:12.964488414 +0000 UTC m=+915.708034665" Dec 10 09:49:13 crc kubenswrapper[4715]: I1210 09:49:13.085815 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" podStartSLOduration=6.692215038 podStartE2EDuration="53.085798521s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.273731798 +0000 UTC m=+866.017278049" lastFinishedPulling="2025-12-10 09:49:09.667315281 +0000 UTC m=+912.410861532" observedRunningTime="2025-12-10 09:49:13.085275647 +0000 UTC m=+915.828821898" watchObservedRunningTime="2025-12-10 09:49:13.085798521 +0000 UTC m=+915.829344772" Dec 10 09:49:13 crc kubenswrapper[4715]: I1210 09:49:13.086936 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" podStartSLOduration=5.926800954 podStartE2EDuration="53.086901361s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:22.763667157 +0000 UTC m=+865.507213408" lastFinishedPulling="2025-12-10 09:49:09.923767564 +0000 UTC m=+912.667313815" observedRunningTime="2025-12-10 09:49:13.053208548 +0000 UTC m=+915.796754829" watchObservedRunningTime="2025-12-10 09:49:13.086901361 +0000 UTC m=+915.830447612" Dec 10 09:49:13 crc kubenswrapper[4715]: I1210 09:49:13.120506 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-x98sw" podStartSLOduration=6.430664158 podStartE2EDuration="53.120489252s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.298767968 +0000 UTC m=+866.042314219" lastFinishedPulling="2025-12-10 09:49:09.988593062 +0000 UTC m=+912.732139313" observedRunningTime="2025-12-10 09:49:13.113325016 +0000 UTC m=+915.856871267" watchObservedRunningTime="2025-12-10 09:49:13.120489252 +0000 UTC m=+915.864035503" Dec 10 09:49:13 crc kubenswrapper[4715]: I1210 09:49:13.152020 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-f8zw7" podStartSLOduration=6.501398085 podStartE2EDuration="53.151998187s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.274006765 +0000 UTC m=+866.017553016" lastFinishedPulling="2025-12-10 09:49:09.924606867 +0000 UTC m=+912.668153118" observedRunningTime="2025-12-10 09:49:13.150182227 +0000 UTC m=+915.893728498" watchObservedRunningTime="2025-12-10 09:49:13.151998187 +0000 UTC m=+915.895544468" Dec 10 09:49:13 crc kubenswrapper[4715]: I1210 09:49:13.175581 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" podStartSLOduration=6.274134301 podStartE2EDuration="53.175566903s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:22.764092219 +0000 UTC m=+865.507638470" lastFinishedPulling="2025-12-10 09:49:09.665524821 +0000 UTC m=+912.409071072" observedRunningTime="2025-12-10 09:49:13.172618262 +0000 UTC m=+915.916164503" watchObservedRunningTime="2025-12-10 09:49:13.175566903 +0000 UTC m=+915.919113154" Dec 10 09:49:13 crc kubenswrapper[4715]: I1210 09:49:13.575498 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" event={"ID":"75ebabe2-8b57-4eee-9b3b-b238fafd94b7","Type":"ContainerStarted","Data":"fdf290652aabdcb449ca8198f3daf6ba5ef9c9268e37552d0a2aa959c3cc35d9"} Dec 10 09:49:13 crc kubenswrapper[4715]: I1210 09:49:13.583547 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" Dec 10 09:49:14 crc kubenswrapper[4715]: I1210 09:49:14.608415 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 09:49:14 crc kubenswrapper[4715]: I1210 09:49:14.626023 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" podStartSLOduration=4.753673914 podStartE2EDuration="54.62600748s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.180749738 +0000 UTC m=+865.924295989" lastFinishedPulling="2025-12-10 09:49:13.053083304 +0000 UTC m=+915.796629555" observedRunningTime="2025-12-10 09:49:13.616600518 +0000 UTC m=+916.360146769" watchObservedRunningTime="2025-12-10 09:49:14.62600748 +0000 UTC m=+917.369553731" Dec 10 09:49:16 crc kubenswrapper[4715]: I1210 09:49:16.608775 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4" event={"ID":"3dccc54c-41e2-4745-839c-df51ca4e1825","Type":"ContainerStarted","Data":"cceadbd27f79982974b9439b9f7039d4f9cddb48c2fa56ec51a573b48ae520d8"} Dec 10 09:49:16 crc kubenswrapper[4715]: I1210 09:49:16.633387 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-69rd4" podStartSLOduration=3.243313516 podStartE2EDuration="55.633371412s" podCreationTimestamp="2025-12-10 09:48:21 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.292415638 +0000 UTC m=+866.035961889" lastFinishedPulling="2025-12-10 09:49:15.682473534 +0000 UTC m=+918.426019785" observedRunningTime="2025-12-10 09:49:16.629787683 +0000 UTC m=+919.373333934" watchObservedRunningTime="2025-12-10 09:49:16.633371412 +0000 UTC m=+919.376917663" Dec 10 09:49:16 crc kubenswrapper[4715]: I1210 09:49:16.761934 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxxcb" Dec 10 09:49:16 crc kubenswrapper[4715]: I1210 09:49:16.986647 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fphth6" Dec 10 09:49:20 crc kubenswrapper[4715]: I1210 09:49:20.779394 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-scbxh" Dec 10 09:49:20 crc kubenswrapper[4715]: I1210 09:49:20.848625 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-d875q" Dec 10 09:49:20 crc kubenswrapper[4715]: I1210 09:49:20.927199 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-h9rpn" Dec 10 09:49:20 crc kubenswrapper[4715]: I1210 09:49:20.952875 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-88z54" Dec 10 09:49:21 crc kubenswrapper[4715]: I1210 09:49:21.094342 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-p69tc" Dec 10 09:49:21 crc kubenswrapper[4715]: I1210 09:49:21.095881 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-tjqbt" Dec 10 09:49:21 crc kubenswrapper[4715]: I1210 09:49:21.156155 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-66fzd" Dec 10 09:49:21 crc kubenswrapper[4715]: I1210 09:49:21.221288 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gbrf4" Dec 10 09:49:21 crc kubenswrapper[4715]: I1210 09:49:21.544458 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-2957k" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.368145 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t6dgd"] Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.369606 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.391772 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t6dgd"] Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.499746 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njknt\" (UniqueName: \"kubernetes.io/projected/4821f1c4-8663-44c0-bbfd-38018aa52d4b-kube-api-access-njknt\") pod \"community-operators-t6dgd\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.499899 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-catalog-content\") pod \"community-operators-t6dgd\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.499983 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-utilities\") pod \"community-operators-t6dgd\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.601549 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-utilities\") pod \"community-operators-t6dgd\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.601844 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njknt\" (UniqueName: \"kubernetes.io/projected/4821f1c4-8663-44c0-bbfd-38018aa52d4b-kube-api-access-njknt\") pod \"community-operators-t6dgd\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.602064 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-catalog-content\") pod \"community-operators-t6dgd\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.602088 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-utilities\") pod \"community-operators-t6dgd\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.602335 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-catalog-content\") pod \"community-operators-t6dgd\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.627898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njknt\" (UniqueName: \"kubernetes.io/projected/4821f1c4-8663-44c0-bbfd-38018aa52d4b-kube-api-access-njknt\") pod \"community-operators-t6dgd\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:22 crc kubenswrapper[4715]: I1210 09:49:22.688179 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:23 crc kubenswrapper[4715]: I1210 09:49:23.261253 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t6dgd"] Dec 10 09:49:23 crc kubenswrapper[4715]: W1210 09:49:23.275388 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4821f1c4_8663_44c0_bbfd_38018aa52d4b.slice/crio-f66a7fcd115a0de4f4412c548ee70ede1c597ae1c055a1ad7904fd2b1484d103 WatchSource:0}: Error finding container f66a7fcd115a0de4f4412c548ee70ede1c597ae1c055a1ad7904fd2b1484d103: Status 404 returned error can't find the container with id f66a7fcd115a0de4f4412c548ee70ede1c597ae1c055a1ad7904fd2b1484d103 Dec 10 09:49:23 crc kubenswrapper[4715]: I1210 09:49:23.518317 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7fd7979c46-c59rw" Dec 10 09:49:23 crc kubenswrapper[4715]: I1210 09:49:23.658162 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6dgd" event={"ID":"4821f1c4-8663-44c0-bbfd-38018aa52d4b","Type":"ContainerStarted","Data":"f66a7fcd115a0de4f4412c548ee70ede1c597ae1c055a1ad7904fd2b1484d103"} Dec 10 09:49:25 crc kubenswrapper[4715]: I1210 09:49:25.676770 4715 generic.go:334] "Generic (PLEG): container finished" podID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerID="2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96" exitCode=0 Dec 10 09:49:25 crc kubenswrapper[4715]: I1210 09:49:25.676845 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6dgd" event={"ID":"4821f1c4-8663-44c0-bbfd-38018aa52d4b","Type":"ContainerDied","Data":"2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96"} Dec 10 09:49:26 crc kubenswrapper[4715]: I1210 09:49:26.685477 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6dgd" event={"ID":"4821f1c4-8663-44c0-bbfd-38018aa52d4b","Type":"ContainerStarted","Data":"ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b"} Dec 10 09:49:27 crc kubenswrapper[4715]: I1210 09:49:27.793462 4715 generic.go:334] "Generic (PLEG): container finished" podID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerID="ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b" exitCode=0 Dec 10 09:49:27 crc kubenswrapper[4715]: I1210 09:49:27.793755 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6dgd" event={"ID":"4821f1c4-8663-44c0-bbfd-38018aa52d4b","Type":"ContainerDied","Data":"ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b"} Dec 10 09:49:27 crc kubenswrapper[4715]: I1210 09:49:27.801355 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" event={"ID":"cfe88ca5-7d1d-4fff-8404-8b192ae48b09","Type":"ContainerStarted","Data":"c83f34054ade048c9450a716bd8af108528d4bf50bb5c78a94c749140045ae62"} Dec 10 09:49:27 crc kubenswrapper[4715]: I1210 09:49:27.802749 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" Dec 10 09:49:27 crc kubenswrapper[4715]: I1210 09:49:27.805517 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" event={"ID":"67b3cab4-c74e-4861-875b-e38aac7ced71","Type":"ContainerStarted","Data":"86b0cace760d346e6fb51f3c00e928685f18a5996650c10e4b1fe893f16d545e"} Dec 10 09:49:27 crc kubenswrapper[4715]: I1210 09:49:27.805757 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" Dec 10 09:49:27 crc kubenswrapper[4715]: I1210 09:49:27.835692 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" podStartSLOduration=4.5606096019999995 podStartE2EDuration="1m7.835674271s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.307157156 +0000 UTC m=+866.050703407" lastFinishedPulling="2025-12-10 09:49:26.582221825 +0000 UTC m=+929.325768076" observedRunningTime="2025-12-10 09:49:27.833266955 +0000 UTC m=+930.576813206" watchObservedRunningTime="2025-12-10 09:49:27.835674271 +0000 UTC m=+930.579220522" Dec 10 09:49:27 crc kubenswrapper[4715]: I1210 09:49:27.853409 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" podStartSLOduration=3.919589743 podStartE2EDuration="1m7.853384116s" podCreationTimestamp="2025-12-10 09:48:20 +0000 UTC" firstStartedPulling="2025-12-10 09:48:23.329020457 +0000 UTC m=+866.072566718" lastFinishedPulling="2025-12-10 09:49:27.26281484 +0000 UTC m=+930.006361091" observedRunningTime="2025-12-10 09:49:27.849045997 +0000 UTC m=+930.592592248" watchObservedRunningTime="2025-12-10 09:49:27.853384116 +0000 UTC m=+930.596930377" Dec 10 09:49:28 crc kubenswrapper[4715]: I1210 09:49:28.813692 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6dgd" event={"ID":"4821f1c4-8663-44c0-bbfd-38018aa52d4b","Type":"ContainerStarted","Data":"a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc"} Dec 10 09:49:28 crc kubenswrapper[4715]: I1210 09:49:28.835905 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t6dgd" podStartSLOduration=4.249094149 podStartE2EDuration="6.83588728s" podCreationTimestamp="2025-12-10 09:49:22 +0000 UTC" firstStartedPulling="2025-12-10 09:49:25.678436769 +0000 UTC m=+928.421983020" lastFinishedPulling="2025-12-10 09:49:28.26522989 +0000 UTC m=+931.008776151" observedRunningTime="2025-12-10 09:49:28.829852375 +0000 UTC m=+931.573398626" watchObservedRunningTime="2025-12-10 09:49:28.83588728 +0000 UTC m=+931.579433531" Dec 10 09:49:31 crc kubenswrapper[4715]: I1210 09:49:31.643597 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wcvvl" Dec 10 09:49:32 crc kubenswrapper[4715]: I1210 09:49:32.689236 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:32 crc kubenswrapper[4715]: I1210 09:49:32.689383 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:32 crc kubenswrapper[4715]: I1210 09:49:32.756116 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:33 crc kubenswrapper[4715]: I1210 09:49:33.889781 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:33 crc kubenswrapper[4715]: I1210 09:49:33.954168 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t6dgd"] Dec 10 09:49:35 crc kubenswrapper[4715]: I1210 09:49:35.856003 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t6dgd" podUID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerName="registry-server" containerID="cri-o://a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc" gracePeriod=2 Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.862838 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.864241 4715 generic.go:334] "Generic (PLEG): container finished" podID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerID="a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc" exitCode=0 Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.864282 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6dgd" event={"ID":"4821f1c4-8663-44c0-bbfd-38018aa52d4b","Type":"ContainerDied","Data":"a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc"} Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.864310 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6dgd" event={"ID":"4821f1c4-8663-44c0-bbfd-38018aa52d4b","Type":"ContainerDied","Data":"f66a7fcd115a0de4f4412c548ee70ede1c597ae1c055a1ad7904fd2b1484d103"} Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.864333 4715 scope.go:117] "RemoveContainer" containerID="a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.884405 4715 scope.go:117] "RemoveContainer" containerID="ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.918239 4715 scope.go:117] "RemoveContainer" containerID="2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.925342 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-utilities\") pod \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.925442 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njknt\" (UniqueName: \"kubernetes.io/projected/4821f1c4-8663-44c0-bbfd-38018aa52d4b-kube-api-access-njknt\") pod \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.925515 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-catalog-content\") pod \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\" (UID: \"4821f1c4-8663-44c0-bbfd-38018aa52d4b\") " Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.926338 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-utilities" (OuterVolumeSpecName: "utilities") pod "4821f1c4-8663-44c0-bbfd-38018aa52d4b" (UID: "4821f1c4-8663-44c0-bbfd-38018aa52d4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.930884 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4821f1c4-8663-44c0-bbfd-38018aa52d4b-kube-api-access-njknt" (OuterVolumeSpecName: "kube-api-access-njknt") pod "4821f1c4-8663-44c0-bbfd-38018aa52d4b" (UID: "4821f1c4-8663-44c0-bbfd-38018aa52d4b"). InnerVolumeSpecName "kube-api-access-njknt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.939620 4715 scope.go:117] "RemoveContainer" containerID="a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc" Dec 10 09:49:36 crc kubenswrapper[4715]: E1210 09:49:36.940636 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc\": container with ID starting with a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc not found: ID does not exist" containerID="a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.940700 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc"} err="failed to get container status \"a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc\": rpc error: code = NotFound desc = could not find container \"a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc\": container with ID starting with a6a1b8b0e67a8ca1ad278bd0236fc7a5d4dde84249e52afd1bd896f5e55d2dbc not found: ID does not exist" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.940733 4715 scope.go:117] "RemoveContainer" containerID="ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b" Dec 10 09:49:36 crc kubenswrapper[4715]: E1210 09:49:36.941132 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b\": container with ID starting with ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b not found: ID does not exist" containerID="ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.941168 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b"} err="failed to get container status \"ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b\": rpc error: code = NotFound desc = could not find container \"ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b\": container with ID starting with ead071b4df5f625df9f92071a34ca5f1b2ff8d1761e9ec94168b05624a7dc34b not found: ID does not exist" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.941193 4715 scope.go:117] "RemoveContainer" containerID="2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96" Dec 10 09:49:36 crc kubenswrapper[4715]: E1210 09:49:36.941560 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96\": container with ID starting with 2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96 not found: ID does not exist" containerID="2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96" Dec 10 09:49:36 crc kubenswrapper[4715]: I1210 09:49:36.941593 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96"} err="failed to get container status \"2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96\": rpc error: code = NotFound desc = could not find container \"2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96\": container with ID starting with 2db6038191c325d1bdece582d0e8b76b89d4f131b4788b432b233b0099a55f96 not found: ID does not exist" Dec 10 09:49:37 crc kubenswrapper[4715]: I1210 09:49:37.025770 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4821f1c4-8663-44c0-bbfd-38018aa52d4b" (UID: "4821f1c4-8663-44c0-bbfd-38018aa52d4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:49:37 crc kubenswrapper[4715]: I1210 09:49:37.027242 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:49:37 crc kubenswrapper[4715]: I1210 09:49:37.027279 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njknt\" (UniqueName: \"kubernetes.io/projected/4821f1c4-8663-44c0-bbfd-38018aa52d4b-kube-api-access-njknt\") on node \"crc\" DevicePath \"\"" Dec 10 09:49:37 crc kubenswrapper[4715]: I1210 09:49:37.027291 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4821f1c4-8663-44c0-bbfd-38018aa52d4b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:49:37 crc kubenswrapper[4715]: I1210 09:49:37.876491 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t6dgd" Dec 10 09:49:37 crc kubenswrapper[4715]: I1210 09:49:37.905224 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t6dgd"] Dec 10 09:49:37 crc kubenswrapper[4715]: I1210 09:49:37.912088 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t6dgd"] Dec 10 09:49:39 crc kubenswrapper[4715]: I1210 09:49:39.613382 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" path="/var/lib/kubelet/pods/4821f1c4-8663-44c0-bbfd-38018aa52d4b/volumes" Dec 10 09:49:41 crc kubenswrapper[4715]: I1210 09:49:41.511074 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-ht6n4" Dec 10 09:49:47 crc kubenswrapper[4715]: I1210 09:49:47.714196 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:49:47 crc kubenswrapper[4715]: I1210 09:49:47.714451 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.601406 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v4sbd"] Dec 10 09:49:57 crc kubenswrapper[4715]: E1210 09:49:57.603437 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerName="extract-utilities" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.603539 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerName="extract-utilities" Dec 10 09:49:57 crc kubenswrapper[4715]: E1210 09:49:57.603740 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerName="registry-server" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.603789 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerName="registry-server" Dec 10 09:49:57 crc kubenswrapper[4715]: E1210 09:49:57.603820 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerName="extract-content" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.603827 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerName="extract-content" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.604150 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4821f1c4-8663-44c0-bbfd-38018aa52d4b" containerName="registry-server" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.605364 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.619393 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4sbd"] Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.708951 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-catalog-content\") pod \"certified-operators-v4sbd\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.709016 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jrth\" (UniqueName: \"kubernetes.io/projected/fc115bed-1bdd-48fa-894a-c1e83329a571-kube-api-access-7jrth\") pod \"certified-operators-v4sbd\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.709077 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-utilities\") pod \"certified-operators-v4sbd\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.809870 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-catalog-content\") pod \"certified-operators-v4sbd\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.810003 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jrth\" (UniqueName: \"kubernetes.io/projected/fc115bed-1bdd-48fa-894a-c1e83329a571-kube-api-access-7jrth\") pod \"certified-operators-v4sbd\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.810092 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-utilities\") pod \"certified-operators-v4sbd\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.810400 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-catalog-content\") pod \"certified-operators-v4sbd\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.810418 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-utilities\") pod \"certified-operators-v4sbd\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.829418 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jrth\" (UniqueName: \"kubernetes.io/projected/fc115bed-1bdd-48fa-894a-c1e83329a571-kube-api-access-7jrth\") pod \"certified-operators-v4sbd\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:57 crc kubenswrapper[4715]: I1210 09:49:57.935211 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:49:58 crc kubenswrapper[4715]: I1210 09:49:58.471464 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v4sbd"] Dec 10 09:49:58 crc kubenswrapper[4715]: W1210 09:49:58.474211 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc115bed_1bdd_48fa_894a_c1e83329a571.slice/crio-3d07db6a4aebfd87357fb699739f0109f2a9c6c2cd3e4a10ccfd3e04c916d052 WatchSource:0}: Error finding container 3d07db6a4aebfd87357fb699739f0109f2a9c6c2cd3e4a10ccfd3e04c916d052: Status 404 returned error can't find the container with id 3d07db6a4aebfd87357fb699739f0109f2a9c6c2cd3e4a10ccfd3e04c916d052 Dec 10 09:49:59 crc kubenswrapper[4715]: I1210 09:49:59.037654 4715 generic.go:334] "Generic (PLEG): container finished" podID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerID="1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496" exitCode=0 Dec 10 09:49:59 crc kubenswrapper[4715]: I1210 09:49:59.037715 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4sbd" event={"ID":"fc115bed-1bdd-48fa-894a-c1e83329a571","Type":"ContainerDied","Data":"1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496"} Dec 10 09:49:59 crc kubenswrapper[4715]: I1210 09:49:59.038040 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4sbd" event={"ID":"fc115bed-1bdd-48fa-894a-c1e83329a571","Type":"ContainerStarted","Data":"3d07db6a4aebfd87357fb699739f0109f2a9c6c2cd3e4a10ccfd3e04c916d052"} Dec 10 09:50:00 crc kubenswrapper[4715]: I1210 09:50:00.048440 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4sbd" event={"ID":"fc115bed-1bdd-48fa-894a-c1e83329a571","Type":"ContainerStarted","Data":"c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699"} Dec 10 09:50:01 crc kubenswrapper[4715]: I1210 09:50:01.062982 4715 generic.go:334] "Generic (PLEG): container finished" podID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerID="c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699" exitCode=0 Dec 10 09:50:01 crc kubenswrapper[4715]: I1210 09:50:01.063033 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4sbd" event={"ID":"fc115bed-1bdd-48fa-894a-c1e83329a571","Type":"ContainerDied","Data":"c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699"} Dec 10 09:50:02 crc kubenswrapper[4715]: I1210 09:50:02.071276 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4sbd" event={"ID":"fc115bed-1bdd-48fa-894a-c1e83329a571","Type":"ContainerStarted","Data":"10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55"} Dec 10 09:50:02 crc kubenswrapper[4715]: I1210 09:50:02.088810 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v4sbd" podStartSLOduration=2.291799172 podStartE2EDuration="5.088789848s" podCreationTimestamp="2025-12-10 09:49:57 +0000 UTC" firstStartedPulling="2025-12-10 09:49:59.039529834 +0000 UTC m=+961.783076085" lastFinishedPulling="2025-12-10 09:50:01.83652051 +0000 UTC m=+964.580066761" observedRunningTime="2025-12-10 09:50:02.087258977 +0000 UTC m=+964.830805228" watchObservedRunningTime="2025-12-10 09:50:02.088789848 +0000 UTC m=+964.832336099" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.547351 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-c5qfr"] Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.554096 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.560990 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-865j8" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.561291 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.561416 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.561443 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.568760 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-c5qfr"] Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.616391 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-b6gh6"] Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.617590 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.620024 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.628566 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-b6gh6"] Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.718707 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkvtg\" (UniqueName: \"kubernetes.io/projected/12c73108-902d-4898-bac5-b196621c1b32-kube-api-access-lkvtg\") pod \"dnsmasq-dns-675f4bcbfc-c5qfr\" (UID: \"12c73108-902d-4898-bac5-b196621c1b32\") " pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.718824 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c73108-902d-4898-bac5-b196621c1b32-config\") pod \"dnsmasq-dns-675f4bcbfc-c5qfr\" (UID: \"12c73108-902d-4898-bac5-b196621c1b32\") " pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.820377 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-b6gh6\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.820420 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c73108-902d-4898-bac5-b196621c1b32-config\") pod \"dnsmasq-dns-675f4bcbfc-c5qfr\" (UID: \"12c73108-902d-4898-bac5-b196621c1b32\") " pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.820476 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkvtg\" (UniqueName: \"kubernetes.io/projected/12c73108-902d-4898-bac5-b196621c1b32-kube-api-access-lkvtg\") pod \"dnsmasq-dns-675f4bcbfc-c5qfr\" (UID: \"12c73108-902d-4898-bac5-b196621c1b32\") " pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.820510 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csk8l\" (UniqueName: \"kubernetes.io/projected/bd00920c-ce65-4cfe-b627-44129f4129dd-kube-api-access-csk8l\") pod \"dnsmasq-dns-78dd6ddcc-b6gh6\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.820586 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-config\") pod \"dnsmasq-dns-78dd6ddcc-b6gh6\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.821448 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c73108-902d-4898-bac5-b196621c1b32-config\") pod \"dnsmasq-dns-675f4bcbfc-c5qfr\" (UID: \"12c73108-902d-4898-bac5-b196621c1b32\") " pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.837527 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkvtg\" (UniqueName: \"kubernetes.io/projected/12c73108-902d-4898-bac5-b196621c1b32-kube-api-access-lkvtg\") pod \"dnsmasq-dns-675f4bcbfc-c5qfr\" (UID: \"12c73108-902d-4898-bac5-b196621c1b32\") " pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.880543 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.924705 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csk8l\" (UniqueName: \"kubernetes.io/projected/bd00920c-ce65-4cfe-b627-44129f4129dd-kube-api-access-csk8l\") pod \"dnsmasq-dns-78dd6ddcc-b6gh6\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.924757 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-config\") pod \"dnsmasq-dns-78dd6ddcc-b6gh6\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.924823 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-b6gh6\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.925797 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-config\") pod \"dnsmasq-dns-78dd6ddcc-b6gh6\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.925975 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-b6gh6\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:05 crc kubenswrapper[4715]: I1210 09:50:05.942224 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csk8l\" (UniqueName: \"kubernetes.io/projected/bd00920c-ce65-4cfe-b627-44129f4129dd-kube-api-access-csk8l\") pod \"dnsmasq-dns-78dd6ddcc-b6gh6\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:06 crc kubenswrapper[4715]: I1210 09:50:06.240705 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:06 crc kubenswrapper[4715]: I1210 09:50:06.563099 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-c5qfr"] Dec 10 09:50:06 crc kubenswrapper[4715]: I1210 09:50:06.723252 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-b6gh6"] Dec 10 09:50:07 crc kubenswrapper[4715]: I1210 09:50:07.106947 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" event={"ID":"bd00920c-ce65-4cfe-b627-44129f4129dd","Type":"ContainerStarted","Data":"d463434a92eb1f17d041b7b2fce5b0aba454306d437bafda4b795528d37e87d3"} Dec 10 09:50:07 crc kubenswrapper[4715]: I1210 09:50:07.108076 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" event={"ID":"12c73108-902d-4898-bac5-b196621c1b32","Type":"ContainerStarted","Data":"85835dc59869102461c4bd1a4e497e7695bd6b38cde914b1996baabbd072e213"} Dec 10 09:50:07 crc kubenswrapper[4715]: I1210 09:50:07.936257 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:50:07 crc kubenswrapper[4715]: I1210 09:50:07.937181 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:50:08 crc kubenswrapper[4715]: I1210 09:50:08.202697 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:50:08 crc kubenswrapper[4715]: I1210 09:50:08.287907 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:50:08 crc kubenswrapper[4715]: I1210 09:50:08.464271 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4sbd"] Dec 10 09:50:08 crc kubenswrapper[4715]: I1210 09:50:08.781950 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-c5qfr"] Dec 10 09:50:08 crc kubenswrapper[4715]: I1210 09:50:08.820141 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dj6bj"] Dec 10 09:50:08 crc kubenswrapper[4715]: I1210 09:50:08.821628 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:08 crc kubenswrapper[4715]: I1210 09:50:08.834853 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dj6bj"] Dec 10 09:50:08 crc kubenswrapper[4715]: I1210 09:50:08.957751 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-config\") pod \"dnsmasq-dns-666b6646f7-dj6bj\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:08 crc kubenswrapper[4715]: I1210 09:50:08.957836 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpdtn\" (UniqueName: \"kubernetes.io/projected/ffb85f8e-eec3-4751-8ff2-31839610dba4-kube-api-access-kpdtn\") pod \"dnsmasq-dns-666b6646f7-dj6bj\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:08 crc kubenswrapper[4715]: I1210 09:50:08.957998 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dj6bj\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.058997 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-config\") pod \"dnsmasq-dns-666b6646f7-dj6bj\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.059049 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpdtn\" (UniqueName: \"kubernetes.io/projected/ffb85f8e-eec3-4751-8ff2-31839610dba4-kube-api-access-kpdtn\") pod \"dnsmasq-dns-666b6646f7-dj6bj\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.059072 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dj6bj\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.060150 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-config\") pod \"dnsmasq-dns-666b6646f7-dj6bj\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.060209 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dj6bj\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.097636 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpdtn\" (UniqueName: \"kubernetes.io/projected/ffb85f8e-eec3-4751-8ff2-31839610dba4-kube-api-access-kpdtn\") pod \"dnsmasq-dns-666b6646f7-dj6bj\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.128945 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-b6gh6"] Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.151987 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lm5b9"] Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.153248 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.170597 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lm5b9"] Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.171179 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.268438 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwh7h\" (UniqueName: \"kubernetes.io/projected/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-kube-api-access-mwh7h\") pod \"dnsmasq-dns-57d769cc4f-lm5b9\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.268501 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-config\") pod \"dnsmasq-dns-57d769cc4f-lm5b9\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.268558 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lm5b9\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.377959 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwh7h\" (UniqueName: \"kubernetes.io/projected/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-kube-api-access-mwh7h\") pod \"dnsmasq-dns-57d769cc4f-lm5b9\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.378261 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-config\") pod \"dnsmasq-dns-57d769cc4f-lm5b9\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.378310 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lm5b9\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.379037 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-config\") pod \"dnsmasq-dns-57d769cc4f-lm5b9\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.379490 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lm5b9\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.419148 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwh7h\" (UniqueName: \"kubernetes.io/projected/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-kube-api-access-mwh7h\") pod \"dnsmasq-dns-57d769cc4f-lm5b9\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.519293 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.960358 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dj6bj"] Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.991688 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 09:50:09 crc kubenswrapper[4715]: I1210 09:50:09.994134 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.001037 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.001207 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.001351 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-85zs9" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.001494 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.003174 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.003297 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.003427 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.021143 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.098868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.098931 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.098981 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-config-data\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.099001 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.099022 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.099044 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwmg5\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-kube-api-access-jwmg5\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.099071 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.099097 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.099118 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.099147 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.099173 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.110997 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lm5b9"] Dec 10 09:50:10 crc kubenswrapper[4715]: W1210 09:50:10.134177 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fbc1385_27ee_4c8b_8ce5_de03f14b4876.slice/crio-7e96b664c205048e259ec4b4af9d546a412e800373f0c95be102623cc23685b2 WatchSource:0}: Error finding container 7e96b664c205048e259ec4b4af9d546a412e800373f0c95be102623cc23685b2: Status 404 returned error can't find the container with id 7e96b664c205048e259ec4b4af9d546a412e800373f0c95be102623cc23685b2 Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.200642 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.200708 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.200738 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.200774 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.200833 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.200854 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.200882 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.200953 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-config-data\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.200982 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.201007 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.201052 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwmg5\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-kube-api-access-jwmg5\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.208344 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.208527 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.208689 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.209329 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-config-data\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.210380 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.211307 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.214498 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" event={"ID":"ffb85f8e-eec3-4751-8ff2-31839610dba4","Type":"ContainerStarted","Data":"60e742c574034e39fbd4b89a33f45a8d6568fcb89a2b7cf82ecd47d143f5fa0f"} Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.225863 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.226321 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" event={"ID":"9fbc1385-27ee-4c8b-8ce5-de03f14b4876","Type":"ContainerStarted","Data":"7e96b664c205048e259ec4b4af9d546a412e800373f0c95be102623cc23685b2"} Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.226494 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v4sbd" podUID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerName="registry-server" containerID="cri-o://10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55" gracePeriod=2 Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.232406 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwmg5\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-kube-api-access-jwmg5\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.244372 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.244413 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.255578 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.294181 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.371022 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.401922 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.413293 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.415798 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.416903 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.419127 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wrvwh" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.419232 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.419282 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.419437 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.419526 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.419674 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.515709 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.515763 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.515845 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.515870 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.515905 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5afc401a-b98d-4b26-8a47-c18582062f49-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.515942 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.515963 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.515988 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.516031 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.516060 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5afc401a-b98d-4b26-8a47-c18582062f49-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.516079 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l2mv\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-kube-api-access-4l2mv\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.619953 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.620023 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.620066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5afc401a-b98d-4b26-8a47-c18582062f49-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.620107 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.620129 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.620154 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.620224 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.620273 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5afc401a-b98d-4b26-8a47-c18582062f49-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.620292 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l2mv\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-kube-api-access-4l2mv\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.620342 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.620367 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.621233 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.622123 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.622144 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.622378 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.622725 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.627390 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.632512 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5afc401a-b98d-4b26-8a47-c18582062f49-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.632695 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.633824 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.648303 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5afc401a-b98d-4b26-8a47-c18582062f49-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.661520 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l2mv\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-kube-api-access-4l2mv\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.675468 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.778419 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.860789 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.924232 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jrth\" (UniqueName: \"kubernetes.io/projected/fc115bed-1bdd-48fa-894a-c1e83329a571-kube-api-access-7jrth\") pod \"fc115bed-1bdd-48fa-894a-c1e83329a571\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.924358 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-utilities\") pod \"fc115bed-1bdd-48fa-894a-c1e83329a571\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.924487 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-catalog-content\") pod \"fc115bed-1bdd-48fa-894a-c1e83329a571\" (UID: \"fc115bed-1bdd-48fa-894a-c1e83329a571\") " Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.925673 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-utilities" (OuterVolumeSpecName: "utilities") pod "fc115bed-1bdd-48fa-894a-c1e83329a571" (UID: "fc115bed-1bdd-48fa-894a-c1e83329a571"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:50:10 crc kubenswrapper[4715]: I1210 09:50:10.928525 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc115bed-1bdd-48fa-894a-c1e83329a571-kube-api-access-7jrth" (OuterVolumeSpecName: "kube-api-access-7jrth") pod "fc115bed-1bdd-48fa-894a-c1e83329a571" (UID: "fc115bed-1bdd-48fa-894a-c1e83329a571"). InnerVolumeSpecName "kube-api-access-7jrth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.026409 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.026448 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jrth\" (UniqueName: \"kubernetes.io/projected/fc115bed-1bdd-48fa-894a-c1e83329a571-kube-api-access-7jrth\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.075173 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 09:50:11 crc kubenswrapper[4715]: W1210 09:50:11.094363 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad27e35f_aceb_42a8_a60f_efd71d0a3b73.slice/crio-d7416f7a399cf0025b65f3f516759f0a43dd20c45091f846fb5ef2446c05debb WatchSource:0}: Error finding container d7416f7a399cf0025b65f3f516759f0a43dd20c45091f846fb5ef2446c05debb: Status 404 returned error can't find the container with id d7416f7a399cf0025b65f3f516759f0a43dd20c45091f846fb5ef2446c05debb Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.117006 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 10 09:50:11 crc kubenswrapper[4715]: E1210 09:50:11.117606 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerName="extract-content" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.117622 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerName="extract-content" Dec 10 09:50:11 crc kubenswrapper[4715]: E1210 09:50:11.117667 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerName="registry-server" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.117674 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerName="registry-server" Dec 10 09:50:11 crc kubenswrapper[4715]: E1210 09:50:11.117684 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerName="extract-utilities" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.117692 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerName="extract-utilities" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.118437 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerName="registry-server" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.119609 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.123995 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.124097 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.124219 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-s64pg" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.125583 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.130987 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.140475 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.230626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-config-data-default\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.230675 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.230721 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.230771 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.230801 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8rwr\" (UniqueName: \"kubernetes.io/projected/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-kube-api-access-g8rwr\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.230823 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.231110 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-kolla-config\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.231242 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.240233 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad27e35f-aceb-42a8-a60f-efd71d0a3b73","Type":"ContainerStarted","Data":"d7416f7a399cf0025b65f3f516759f0a43dd20c45091f846fb5ef2446c05debb"} Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.242940 4715 generic.go:334] "Generic (PLEG): container finished" podID="fc115bed-1bdd-48fa-894a-c1e83329a571" containerID="10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55" exitCode=0 Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.243001 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4sbd" event={"ID":"fc115bed-1bdd-48fa-894a-c1e83329a571","Type":"ContainerDied","Data":"10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55"} Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.243021 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v4sbd" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.243045 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v4sbd" event={"ID":"fc115bed-1bdd-48fa-894a-c1e83329a571","Type":"ContainerDied","Data":"3d07db6a4aebfd87357fb699739f0109f2a9c6c2cd3e4a10ccfd3e04c916d052"} Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.243068 4715 scope.go:117] "RemoveContainer" containerID="10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.289407 4715 scope.go:117] "RemoveContainer" containerID="c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.312126 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.325027 4715 scope.go:117] "RemoveContainer" containerID="1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.332450 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-config-data-default\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.332485 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.332526 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.332627 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.332661 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8rwr\" (UniqueName: \"kubernetes.io/projected/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-kube-api-access-g8rwr\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.332683 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.332713 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-kolla-config\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.332736 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.332935 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.333325 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.333540 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-kolla-config\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.333851 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-config-data-default\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.338653 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.340533 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.345143 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.351674 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8rwr\" (UniqueName: \"kubernetes.io/projected/8bcd31bd-4e40-41b6-92ef-6e0e0b170afa-kube-api-access-g8rwr\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.362738 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa\") " pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.418901 4715 scope.go:117] "RemoveContainer" containerID="10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55" Dec 10 09:50:11 crc kubenswrapper[4715]: E1210 09:50:11.419365 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55\": container with ID starting with 10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55 not found: ID does not exist" containerID="10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.419397 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55"} err="failed to get container status \"10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55\": rpc error: code = NotFound desc = could not find container \"10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55\": container with ID starting with 10f1a1dff43a7ca5e523d323f9ed4ecf6652f81ae73761019f3218b9153d2b55 not found: ID does not exist" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.419421 4715 scope.go:117] "RemoveContainer" containerID="c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699" Dec 10 09:50:11 crc kubenswrapper[4715]: E1210 09:50:11.424191 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699\": container with ID starting with c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699 not found: ID does not exist" containerID="c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.424249 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699"} err="failed to get container status \"c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699\": rpc error: code = NotFound desc = could not find container \"c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699\": container with ID starting with c41a6d31cad75ba6077927d74b9daefebfcdf3b5a9052ed00b66c9544ca95699 not found: ID does not exist" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.424276 4715 scope.go:117] "RemoveContainer" containerID="1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496" Dec 10 09:50:11 crc kubenswrapper[4715]: E1210 09:50:11.427468 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496\": container with ID starting with 1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496 not found: ID does not exist" containerID="1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.429339 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496"} err="failed to get container status \"1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496\": rpc error: code = NotFound desc = could not find container \"1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496\": container with ID starting with 1f366cf957d54efe1dfa73829d3bb0c47f028ac86262df858cedc3d7dc047496 not found: ID does not exist" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.446571 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc115bed-1bdd-48fa-894a-c1e83329a571" (UID: "fc115bed-1bdd-48fa-894a-c1e83329a571"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.466650 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.537397 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc115bed-1bdd-48fa-894a-c1e83329a571-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.659145 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v4sbd"] Dec 10 09:50:11 crc kubenswrapper[4715]: I1210 09:50:11.670980 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v4sbd"] Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.028612 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.260396 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa","Type":"ContainerStarted","Data":"212f42273182f5dc8af5d68370f8ed7b788ccaf3126f3dc367754dad14f46d56"} Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.262640 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5afc401a-b98d-4b26-8a47-c18582062f49","Type":"ContainerStarted","Data":"4a81c8b991df4ee7ce5500dc6e43bf36260001444db6237d0ca7fe62778e813c"} Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.677523 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.680430 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.681877 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.684306 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-5cc6l" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.684964 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.685102 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.685270 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.773260 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.773334 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/05dd65e3-40ea-4d12-856f-3fa4f99712b2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.773374 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05dd65e3-40ea-4d12-856f-3fa4f99712b2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.773399 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfjjt\" (UniqueName: \"kubernetes.io/projected/05dd65e3-40ea-4d12-856f-3fa4f99712b2-kube-api-access-nfjjt\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.773437 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/05dd65e3-40ea-4d12-856f-3fa4f99712b2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.773456 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05dd65e3-40ea-4d12-856f-3fa4f99712b2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.773483 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/05dd65e3-40ea-4d12-856f-3fa4f99712b2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.773507 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/05dd65e3-40ea-4d12-856f-3fa4f99712b2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.875088 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.875349 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/05dd65e3-40ea-4d12-856f-3fa4f99712b2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.875482 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05dd65e3-40ea-4d12-856f-3fa4f99712b2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.875562 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfjjt\" (UniqueName: \"kubernetes.io/projected/05dd65e3-40ea-4d12-856f-3fa4f99712b2-kube-api-access-nfjjt\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.875491 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.875735 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/05dd65e3-40ea-4d12-856f-3fa4f99712b2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.875828 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05dd65e3-40ea-4d12-856f-3fa4f99712b2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.875948 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/05dd65e3-40ea-4d12-856f-3fa4f99712b2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.876043 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/05dd65e3-40ea-4d12-856f-3fa4f99712b2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.876541 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/05dd65e3-40ea-4d12-856f-3fa4f99712b2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.876570 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/05dd65e3-40ea-4d12-856f-3fa4f99712b2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.877018 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05dd65e3-40ea-4d12-856f-3fa4f99712b2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.877252 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/05dd65e3-40ea-4d12-856f-3fa4f99712b2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.882169 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/05dd65e3-40ea-4d12-856f-3fa4f99712b2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.899008 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfjjt\" (UniqueName: \"kubernetes.io/projected/05dd65e3-40ea-4d12-856f-3fa4f99712b2-kube-api-access-nfjjt\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.907083 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05dd65e3-40ea-4d12-856f-3fa4f99712b2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.917810 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"05dd65e3-40ea-4d12-856f-3fa4f99712b2\") " pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.988215 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.989206 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.990586 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.993224 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 10 09:50:12 crc kubenswrapper[4715]: I1210 09:50:12.994772 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-2cfj2" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.002401 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.009068 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.079834 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92be40ca-d175-4ec0-819b-1579787701d7-config-data\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.080398 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbmmr\" (UniqueName: \"kubernetes.io/projected/92be40ca-d175-4ec0-819b-1579787701d7-kube-api-access-jbmmr\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.080468 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92be40ca-d175-4ec0-819b-1579787701d7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.080704 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/92be40ca-d175-4ec0-819b-1579787701d7-kolla-config\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.080759 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/92be40ca-d175-4ec0-819b-1579787701d7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.184155 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/92be40ca-d175-4ec0-819b-1579787701d7-kolla-config\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.184441 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/92be40ca-d175-4ec0-819b-1579787701d7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.184502 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92be40ca-d175-4ec0-819b-1579787701d7-config-data\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.184538 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbmmr\" (UniqueName: \"kubernetes.io/projected/92be40ca-d175-4ec0-819b-1579787701d7-kube-api-access-jbmmr\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.184562 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92be40ca-d175-4ec0-819b-1579787701d7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.185098 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/92be40ca-d175-4ec0-819b-1579787701d7-kolla-config\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.185764 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92be40ca-d175-4ec0-819b-1579787701d7-config-data\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.204789 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/92be40ca-d175-4ec0-819b-1579787701d7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.218152 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92be40ca-d175-4ec0-819b-1579787701d7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.223274 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbmmr\" (UniqueName: \"kubernetes.io/projected/92be40ca-d175-4ec0-819b-1579787701d7-kube-api-access-jbmmr\") pod \"memcached-0\" (UID: \"92be40ca-d175-4ec0-819b-1579787701d7\") " pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.334972 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.596285 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.618120 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc115bed-1bdd-48fa-894a-c1e83329a571" path="/var/lib/kubelet/pods/fc115bed-1bdd-48fa-894a-c1e83329a571/volumes" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.861015 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gclpm"] Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.863159 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.869837 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.877520 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gclpm"] Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.896383 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-utilities\") pod \"redhat-marketplace-gclpm\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.896436 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-catalog-content\") pod \"redhat-marketplace-gclpm\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:13 crc kubenswrapper[4715]: I1210 09:50:13.896539 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7mkc\" (UniqueName: \"kubernetes.io/projected/f65cc319-a237-48f7-a12a-96a923cfbea6-kube-api-access-g7mkc\") pod \"redhat-marketplace-gclpm\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:13.998116 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7mkc\" (UniqueName: \"kubernetes.io/projected/f65cc319-a237-48f7-a12a-96a923cfbea6-kube-api-access-g7mkc\") pod \"redhat-marketplace-gclpm\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:13.998250 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-utilities\") pod \"redhat-marketplace-gclpm\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:13.998288 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-catalog-content\") pod \"redhat-marketplace-gclpm\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:13.999166 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-utilities\") pod \"redhat-marketplace-gclpm\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:13.999722 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-catalog-content\") pod \"redhat-marketplace-gclpm\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:14.027517 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7mkc\" (UniqueName: \"kubernetes.io/projected/f65cc319-a237-48f7-a12a-96a923cfbea6-kube-api-access-g7mkc\") pod \"redhat-marketplace-gclpm\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:14.198184 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:14.289374 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"05dd65e3-40ea-4d12-856f-3fa4f99712b2","Type":"ContainerStarted","Data":"4224ce89970f88687316301a3f4d37d521d0d78b3d642d8b93290683085b6cfb"} Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:14.925052 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:14.927679 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:14.931413 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-5p8wq" Dec 10 09:50:14 crc kubenswrapper[4715]: I1210 09:50:14.944795 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 09:50:15 crc kubenswrapper[4715]: I1210 09:50:15.042655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q55hp\" (UniqueName: \"kubernetes.io/projected/66bffaf0-4efe-435f-8092-a14c323cb97d-kube-api-access-q55hp\") pod \"kube-state-metrics-0\" (UID: \"66bffaf0-4efe-435f-8092-a14c323cb97d\") " pod="openstack/kube-state-metrics-0" Dec 10 09:50:15 crc kubenswrapper[4715]: I1210 09:50:15.144310 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q55hp\" (UniqueName: \"kubernetes.io/projected/66bffaf0-4efe-435f-8092-a14c323cb97d-kube-api-access-q55hp\") pod \"kube-state-metrics-0\" (UID: \"66bffaf0-4efe-435f-8092-a14c323cb97d\") " pod="openstack/kube-state-metrics-0" Dec 10 09:50:15 crc kubenswrapper[4715]: I1210 09:50:15.175231 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q55hp\" (UniqueName: \"kubernetes.io/projected/66bffaf0-4efe-435f-8092-a14c323cb97d-kube-api-access-q55hp\") pod \"kube-state-metrics-0\" (UID: \"66bffaf0-4efe-435f-8092-a14c323cb97d\") " pod="openstack/kube-state-metrics-0" Dec 10 09:50:15 crc kubenswrapper[4715]: I1210 09:50:15.263297 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 09:50:17 crc kubenswrapper[4715]: I1210 09:50:17.714254 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:50:17 crc kubenswrapper[4715]: I1210 09:50:17.714502 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.581499 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-lz7jm"] Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.582836 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.588986 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.589139 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.589575 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-48z2w" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.589935 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-lz7jm"] Dec 10 09:50:18 crc kubenswrapper[4715]: W1210 09:50:18.602090 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92be40ca_d175_4ec0_819b_1579787701d7.slice/crio-7473c55b87292b40f12b7d9ab6b615c8aaff2dd85aa8f4eb0acb6348cf64a146 WatchSource:0}: Error finding container 7473c55b87292b40f12b7d9ab6b615c8aaff2dd85aa8f4eb0acb6348cf64a146: Status 404 returned error can't find the container with id 7473c55b87292b40f12b7d9ab6b615c8aaff2dd85aa8f4eb0acb6348cf64a146 Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.653980 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-rv2hq"] Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.655933 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.671755 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rv2hq"] Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.720850 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6fc17771-af47-4ba4-be8a-6b3143a999f4-var-log-ovn\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.720978 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-var-run\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721007 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-var-lib\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721040 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-etc-ovs\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721059 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6fc17771-af47-4ba4-be8a-6b3143a999f4-scripts\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721088 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vth6j\" (UniqueName: \"kubernetes.io/projected/6fc17771-af47-4ba4-be8a-6b3143a999f4-kube-api-access-vth6j\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721120 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6fc17771-af47-4ba4-be8a-6b3143a999f4-var-run-ovn\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721143 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm64s\" (UniqueName: \"kubernetes.io/projected/d6ae191b-a288-4185-b98b-15654b1e5dbf-kube-api-access-cm64s\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721168 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6fc17771-af47-4ba4-be8a-6b3143a999f4-var-run\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721193 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fc17771-af47-4ba4-be8a-6b3143a999f4-combined-ca-bundle\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721251 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-var-log\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721276 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fc17771-af47-4ba4-be8a-6b3143a999f4-ovn-controller-tls-certs\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.721306 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6ae191b-a288-4185-b98b-15654b1e5dbf-scripts\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.822944 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6fc17771-af47-4ba4-be8a-6b3143a999f4-var-log-ovn\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823045 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-var-run\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823069 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-var-lib\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823105 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-etc-ovs\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823127 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6fc17771-af47-4ba4-be8a-6b3143a999f4-scripts\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823164 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vth6j\" (UniqueName: \"kubernetes.io/projected/6fc17771-af47-4ba4-be8a-6b3143a999f4-kube-api-access-vth6j\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823222 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6fc17771-af47-4ba4-be8a-6b3143a999f4-var-run-ovn\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823251 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm64s\" (UniqueName: \"kubernetes.io/projected/d6ae191b-a288-4185-b98b-15654b1e5dbf-kube-api-access-cm64s\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823275 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6fc17771-af47-4ba4-be8a-6b3143a999f4-var-run\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823302 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fc17771-af47-4ba4-be8a-6b3143a999f4-combined-ca-bundle\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823333 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-var-log\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823361 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fc17771-af47-4ba4-be8a-6b3143a999f4-ovn-controller-tls-certs\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823390 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6ae191b-a288-4185-b98b-15654b1e5dbf-scripts\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823529 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6fc17771-af47-4ba4-be8a-6b3143a999f4-var-log-ovn\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823633 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-var-lib\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823738 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6fc17771-af47-4ba4-be8a-6b3143a999f4-var-run\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823762 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-var-run\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823786 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-var-log\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823857 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6fc17771-af47-4ba4-be8a-6b3143a999f4-var-run-ovn\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.823884 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d6ae191b-a288-4185-b98b-15654b1e5dbf-etc-ovs\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.827857 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6fc17771-af47-4ba4-be8a-6b3143a999f4-scripts\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.830850 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fc17771-af47-4ba4-be8a-6b3143a999f4-ovn-controller-tls-certs\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.831022 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fc17771-af47-4ba4-be8a-6b3143a999f4-combined-ca-bundle\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.833352 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6ae191b-a288-4185-b98b-15654b1e5dbf-scripts\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.843745 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm64s\" (UniqueName: \"kubernetes.io/projected/d6ae191b-a288-4185-b98b-15654b1e5dbf-kube-api-access-cm64s\") pod \"ovn-controller-ovs-rv2hq\" (UID: \"d6ae191b-a288-4185-b98b-15654b1e5dbf\") " pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.847905 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vth6j\" (UniqueName: \"kubernetes.io/projected/6fc17771-af47-4ba4-be8a-6b3143a999f4-kube-api-access-vth6j\") pod \"ovn-controller-lz7jm\" (UID: \"6fc17771-af47-4ba4-be8a-6b3143a999f4\") " pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.949265 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:18 crc kubenswrapper[4715]: I1210 09:50:18.975854 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.097437 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.098867 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.100738 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.100756 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.100823 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.101633 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-7wsj7" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.101652 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.103724 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.230825 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b4bfd970-df85-4d59-a211-ba4adbb501ee-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.231091 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4bfd970-df85-4d59-a211-ba4adbb501ee-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.231166 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4bfd970-df85-4d59-a211-ba4adbb501ee-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.231278 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bfd970-df85-4d59-a211-ba4adbb501ee-config\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.231373 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.231540 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpt29\" (UniqueName: \"kubernetes.io/projected/b4bfd970-df85-4d59-a211-ba4adbb501ee-kube-api-access-vpt29\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.231795 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4bfd970-df85-4d59-a211-ba4adbb501ee-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.231882 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4bfd970-df85-4d59-a211-ba4adbb501ee-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.333611 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b4bfd970-df85-4d59-a211-ba4adbb501ee-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.333869 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4bfd970-df85-4d59-a211-ba4adbb501ee-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.334024 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4bfd970-df85-4d59-a211-ba4adbb501ee-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.334109 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bfd970-df85-4d59-a211-ba4adbb501ee-config\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.334206 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b4bfd970-df85-4d59-a211-ba4adbb501ee-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.334227 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.334420 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpt29\" (UniqueName: \"kubernetes.io/projected/b4bfd970-df85-4d59-a211-ba4adbb501ee-kube-api-access-vpt29\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.334576 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4bfd970-df85-4d59-a211-ba4adbb501ee-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.334671 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4bfd970-df85-4d59-a211-ba4adbb501ee-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.334993 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.335310 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4bfd970-df85-4d59-a211-ba4adbb501ee-config\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.336410 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4bfd970-df85-4d59-a211-ba4adbb501ee-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.337390 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"92be40ca-d175-4ec0-819b-1579787701d7","Type":"ContainerStarted","Data":"7473c55b87292b40f12b7d9ab6b615c8aaff2dd85aa8f4eb0acb6348cf64a146"} Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.338379 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4bfd970-df85-4d59-a211-ba4adbb501ee-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.339955 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4bfd970-df85-4d59-a211-ba4adbb501ee-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.341083 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4bfd970-df85-4d59-a211-ba4adbb501ee-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.354981 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpt29\" (UniqueName: \"kubernetes.io/projected/b4bfd970-df85-4d59-a211-ba4adbb501ee-kube-api-access-vpt29\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.364682 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b4bfd970-df85-4d59-a211-ba4adbb501ee\") " pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:19 crc kubenswrapper[4715]: I1210 09:50:19.420822 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.213231 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.217356 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.219284 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.219603 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.219736 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-lgxn4" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.220997 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.227289 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.405641 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqsrr\" (UniqueName: \"kubernetes.io/projected/219b8c44-8f39-40f1-9000-25404c35d495-kube-api-access-mqsrr\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.405720 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/219b8c44-8f39-40f1-9000-25404c35d495-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.405763 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/219b8c44-8f39-40f1-9000-25404c35d495-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.405824 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.405851 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/219b8c44-8f39-40f1-9000-25404c35d495-config\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.405958 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/219b8c44-8f39-40f1-9000-25404c35d495-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.406021 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219b8c44-8f39-40f1-9000-25404c35d495-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.406044 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/219b8c44-8f39-40f1-9000-25404c35d495-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.507617 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/219b8c44-8f39-40f1-9000-25404c35d495-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.508792 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219b8c44-8f39-40f1-9000-25404c35d495-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.508852 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/219b8c44-8f39-40f1-9000-25404c35d495-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.508899 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqsrr\" (UniqueName: \"kubernetes.io/projected/219b8c44-8f39-40f1-9000-25404c35d495-kube-api-access-mqsrr\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.508957 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/219b8c44-8f39-40f1-9000-25404c35d495-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.508990 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/219b8c44-8f39-40f1-9000-25404c35d495-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.509026 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.509059 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/219b8c44-8f39-40f1-9000-25404c35d495-config\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.510181 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.510880 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/219b8c44-8f39-40f1-9000-25404c35d495-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.511005 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/219b8c44-8f39-40f1-9000-25404c35d495-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.512333 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/219b8c44-8f39-40f1-9000-25404c35d495-config\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.514264 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219b8c44-8f39-40f1-9000-25404c35d495-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.514504 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/219b8c44-8f39-40f1-9000-25404c35d495-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.519675 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/219b8c44-8f39-40f1-9000-25404c35d495-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.531358 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqsrr\" (UniqueName: \"kubernetes.io/projected/219b8c44-8f39-40f1-9000-25404c35d495-kube-api-access-mqsrr\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.551157 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"219b8c44-8f39-40f1-9000-25404c35d495\") " pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:22 crc kubenswrapper[4715]: I1210 09:50:22.843505 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 10 09:50:30 crc kubenswrapper[4715]: E1210 09:50:30.664989 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 09:50:30 crc kubenswrapper[4715]: E1210 09:50:30.665686 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kpdtn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-dj6bj_openstack(ffb85f8e-eec3-4751-8ff2-31839610dba4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:50:30 crc kubenswrapper[4715]: E1210 09:50:30.668027 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" podUID="ffb85f8e-eec3-4751-8ff2-31839610dba4" Dec 10 09:50:31 crc kubenswrapper[4715]: E1210 09:50:31.460032 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" podUID="ffb85f8e-eec3-4751-8ff2-31839610dba4" Dec 10 09:50:32 crc kubenswrapper[4715]: E1210 09:50:32.400017 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 09:50:32 crc kubenswrapper[4715]: E1210 09:50:32.400484 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-csk8l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-b6gh6_openstack(bd00920c-ce65-4cfe-b627-44129f4129dd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:50:32 crc kubenswrapper[4715]: E1210 09:50:32.401706 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" podUID="bd00920c-ce65-4cfe-b627-44129f4129dd" Dec 10 09:50:32 crc kubenswrapper[4715]: E1210 09:50:32.405865 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 09:50:32 crc kubenswrapper[4715]: E1210 09:50:32.406094 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lkvtg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-c5qfr_openstack(12c73108-902d-4898-bac5-b196621c1b32): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:50:32 crc kubenswrapper[4715]: E1210 09:50:32.408017 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" podUID="12c73108-902d-4898-bac5-b196621c1b32" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.230196 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.249492 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.292568 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c73108-902d-4898-bac5-b196621c1b32-config\") pod \"12c73108-902d-4898-bac5-b196621c1b32\" (UID: \"12c73108-902d-4898-bac5-b196621c1b32\") " Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.292766 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkvtg\" (UniqueName: \"kubernetes.io/projected/12c73108-902d-4898-bac5-b196621c1b32-kube-api-access-lkvtg\") pod \"12c73108-902d-4898-bac5-b196621c1b32\" (UID: \"12c73108-902d-4898-bac5-b196621c1b32\") " Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.293619 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12c73108-902d-4898-bac5-b196621c1b32-config" (OuterVolumeSpecName: "config") pod "12c73108-902d-4898-bac5-b196621c1b32" (UID: "12c73108-902d-4898-bac5-b196621c1b32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.316954 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c73108-902d-4898-bac5-b196621c1b32-kube-api-access-lkvtg" (OuterVolumeSpecName: "kube-api-access-lkvtg") pod "12c73108-902d-4898-bac5-b196621c1b32" (UID: "12c73108-902d-4898-bac5-b196621c1b32"). InnerVolumeSpecName "kube-api-access-lkvtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.394415 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-dns-svc\") pod \"bd00920c-ce65-4cfe-b627-44129f4129dd\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.394990 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csk8l\" (UniqueName: \"kubernetes.io/projected/bd00920c-ce65-4cfe-b627-44129f4129dd-kube-api-access-csk8l\") pod \"bd00920c-ce65-4cfe-b627-44129f4129dd\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.395021 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-config\") pod \"bd00920c-ce65-4cfe-b627-44129f4129dd\" (UID: \"bd00920c-ce65-4cfe-b627-44129f4129dd\") " Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.395062 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd00920c-ce65-4cfe-b627-44129f4129dd" (UID: "bd00920c-ce65-4cfe-b627-44129f4129dd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.395700 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-config" (OuterVolumeSpecName: "config") pod "bd00920c-ce65-4cfe-b627-44129f4129dd" (UID: "bd00920c-ce65-4cfe-b627-44129f4129dd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.396289 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.396336 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkvtg\" (UniqueName: \"kubernetes.io/projected/12c73108-902d-4898-bac5-b196621c1b32-kube-api-access-lkvtg\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.396350 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd00920c-ce65-4cfe-b627-44129f4129dd-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.396360 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c73108-902d-4898-bac5-b196621c1b32-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.466587 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd00920c-ce65-4cfe-b627-44129f4129dd-kube-api-access-csk8l" (OuterVolumeSpecName: "kube-api-access-csk8l") pod "bd00920c-ce65-4cfe-b627-44129f4129dd" (UID: "bd00920c-ce65-4cfe-b627-44129f4129dd"). InnerVolumeSpecName "kube-api-access-csk8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.484978 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" event={"ID":"12c73108-902d-4898-bac5-b196621c1b32","Type":"ContainerDied","Data":"85835dc59869102461c4bd1a4e497e7695bd6b38cde914b1996baabbd072e213"} Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.485613 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-c5qfr" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.494259 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" event={"ID":"bd00920c-ce65-4cfe-b627-44129f4129dd","Type":"ContainerDied","Data":"d463434a92eb1f17d041b7b2fce5b0aba454306d437bafda4b795528d37e87d3"} Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.494357 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-b6gh6" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.504403 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csk8l\" (UniqueName: \"kubernetes.io/projected/bd00920c-ce65-4cfe-b627-44129f4129dd-kube-api-access-csk8l\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.655288 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gclpm"] Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.791512 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-c5qfr"] Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.826106 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-c5qfr"] Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.848019 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-b6gh6"] Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.853841 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-b6gh6"] Dec 10 09:50:34 crc kubenswrapper[4715]: I1210 09:50:34.898719 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.116942 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 09:50:35 crc kubenswrapper[4715]: W1210 09:50:35.152993 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4bfd970_df85_4d59_a211_ba4adbb501ee.slice/crio-099ea97b75749e58b20bb9c4edd8c9b4f726d684c1a4b11f09415acb2fc694f1 WatchSource:0}: Error finding container 099ea97b75749e58b20bb9c4edd8c9b4f726d684c1a4b11f09415acb2fc694f1: Status 404 returned error can't find the container with id 099ea97b75749e58b20bb9c4edd8c9b4f726d684c1a4b11f09415acb2fc694f1 Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.161799 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-lz7jm"] Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.217948 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 09:50:35 crc kubenswrapper[4715]: W1210 09:50:35.248391 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fc17771_af47_4ba4_be8a_6b3143a999f4.slice/crio-2377e70cec0b8acba82708e96692b51e2228cda44a4c85cb04e087d2d4e81529 WatchSource:0}: Error finding container 2377e70cec0b8acba82708e96692b51e2228cda44a4c85cb04e087d2d4e81529: Status 404 returned error can't find the container with id 2377e70cec0b8acba82708e96692b51e2228cda44a4c85cb04e087d2d4e81529 Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.509200 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"219b8c44-8f39-40f1-9000-25404c35d495","Type":"ContainerStarted","Data":"3dbe86898cdb04153aaa22316f937d6b2b42477310ed7cee3a6c0e7f6b4e52e1"} Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.512899 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-lz7jm" event={"ID":"6fc17771-af47-4ba4-be8a-6b3143a999f4","Type":"ContainerStarted","Data":"2377e70cec0b8acba82708e96692b51e2228cda44a4c85cb04e087d2d4e81529"} Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.514276 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b4bfd970-df85-4d59-a211-ba4adbb501ee","Type":"ContainerStarted","Data":"099ea97b75749e58b20bb9c4edd8c9b4f726d684c1a4b11f09415acb2fc694f1"} Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.515944 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"92be40ca-d175-4ec0-819b-1579787701d7","Type":"ContainerStarted","Data":"e3b456731bee622d2b63a55b579fc06461cbc5b2fdd8681dd93752e512980565"} Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.516070 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.517377 4715 generic.go:334] "Generic (PLEG): container finished" podID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerID="4f0bfe2af2c33be5743d4e13bb9741870c9a9e12f4b24a331f20d4e415940f19" exitCode=0 Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.517428 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gclpm" event={"ID":"f65cc319-a237-48f7-a12a-96a923cfbea6","Type":"ContainerDied","Data":"4f0bfe2af2c33be5743d4e13bb9741870c9a9e12f4b24a331f20d4e415940f19"} Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.517445 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gclpm" event={"ID":"f65cc319-a237-48f7-a12a-96a923cfbea6","Type":"ContainerStarted","Data":"af2dc2732d9ad0b09cc98331b5540c5dbf3ca20c3e5835b1d0231e388edfb4a9"} Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.519077 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa","Type":"ContainerStarted","Data":"993e74168ab63e1af0a2cf10ab211c59100add47b309cd832d31bc00532ec3fc"} Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.524648 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66bffaf0-4efe-435f-8092-a14c323cb97d","Type":"ContainerStarted","Data":"0aec39697e8f2909b71d014aa11aa278abb71385d92caf441026b4467852c635"} Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.526303 4715 generic.go:334] "Generic (PLEG): container finished" podID="9fbc1385-27ee-4c8b-8ce5-de03f14b4876" containerID="5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5" exitCode=0 Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.526348 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" event={"ID":"9fbc1385-27ee-4c8b-8ce5-de03f14b4876","Type":"ContainerDied","Data":"5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5"} Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.527610 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"05dd65e3-40ea-4d12-856f-3fa4f99712b2","Type":"ContainerStarted","Data":"08e83a913dc6fc2c55e45154cc6ab39f9fd34e02a83bec96e5f874d4d85f996b"} Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.539803 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=7.834802406 podStartE2EDuration="23.53978017s" podCreationTimestamp="2025-12-10 09:50:12 +0000 UTC" firstStartedPulling="2025-12-10 09:50:18.608585407 +0000 UTC m=+981.352131648" lastFinishedPulling="2025-12-10 09:50:34.313563161 +0000 UTC m=+997.057109412" observedRunningTime="2025-12-10 09:50:35.535168883 +0000 UTC m=+998.278715144" watchObservedRunningTime="2025-12-10 09:50:35.53978017 +0000 UTC m=+998.283326421" Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.616542 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c73108-902d-4898-bac5-b196621c1b32" path="/var/lib/kubelet/pods/12c73108-902d-4898-bac5-b196621c1b32/volumes" Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.617242 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd00920c-ce65-4cfe-b627-44129f4129dd" path="/var/lib/kubelet/pods/bd00920c-ce65-4cfe-b627-44129f4129dd/volumes" Dec 10 09:50:35 crc kubenswrapper[4715]: I1210 09:50:35.901405 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rv2hq"] Dec 10 09:50:36 crc kubenswrapper[4715]: W1210 09:50:36.047488 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6ae191b_a288_4185_b98b_15654b1e5dbf.slice/crio-646e845ff24fc189c225312f6fca335378cb7c214123d63f055ea6f1f2f0e9a2 WatchSource:0}: Error finding container 646e845ff24fc189c225312f6fca335378cb7c214123d63f055ea6f1f2f0e9a2: Status 404 returned error can't find the container with id 646e845ff24fc189c225312f6fca335378cb7c214123d63f055ea6f1f2f0e9a2 Dec 10 09:50:36 crc kubenswrapper[4715]: I1210 09:50:36.537510 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rv2hq" event={"ID":"d6ae191b-a288-4185-b98b-15654b1e5dbf","Type":"ContainerStarted","Data":"646e845ff24fc189c225312f6fca335378cb7c214123d63f055ea6f1f2f0e9a2"} Dec 10 09:50:36 crc kubenswrapper[4715]: I1210 09:50:36.539430 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad27e35f-aceb-42a8-a60f-efd71d0a3b73","Type":"ContainerStarted","Data":"ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2"} Dec 10 09:50:36 crc kubenswrapper[4715]: I1210 09:50:36.544946 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" event={"ID":"9fbc1385-27ee-4c8b-8ce5-de03f14b4876","Type":"ContainerStarted","Data":"fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692"} Dec 10 09:50:36 crc kubenswrapper[4715]: I1210 09:50:36.545036 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:36 crc kubenswrapper[4715]: I1210 09:50:36.547219 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5afc401a-b98d-4b26-8a47-c18582062f49","Type":"ContainerStarted","Data":"993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c"} Dec 10 09:50:36 crc kubenswrapper[4715]: I1210 09:50:36.623389 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" podStartSLOduration=3.480850687 podStartE2EDuration="27.623371356s" podCreationTimestamp="2025-12-10 09:50:09 +0000 UTC" firstStartedPulling="2025-12-10 09:50:10.137405529 +0000 UTC m=+972.880951780" lastFinishedPulling="2025-12-10 09:50:34.279926188 +0000 UTC m=+997.023472449" observedRunningTime="2025-12-10 09:50:36.62242668 +0000 UTC m=+999.365972931" watchObservedRunningTime="2025-12-10 09:50:36.623371356 +0000 UTC m=+999.366917607" Dec 10 09:50:39 crc kubenswrapper[4715]: I1210 09:50:39.569498 4715 generic.go:334] "Generic (PLEG): container finished" podID="8bcd31bd-4e40-41b6-92ef-6e0e0b170afa" containerID="993e74168ab63e1af0a2cf10ab211c59100add47b309cd832d31bc00532ec3fc" exitCode=0 Dec 10 09:50:39 crc kubenswrapper[4715]: I1210 09:50:39.569583 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa","Type":"ContainerDied","Data":"993e74168ab63e1af0a2cf10ab211c59100add47b309cd832d31bc00532ec3fc"} Dec 10 09:50:39 crc kubenswrapper[4715]: I1210 09:50:39.573489 4715 generic.go:334] "Generic (PLEG): container finished" podID="05dd65e3-40ea-4d12-856f-3fa4f99712b2" containerID="08e83a913dc6fc2c55e45154cc6ab39f9fd34e02a83bec96e5f874d4d85f996b" exitCode=0 Dec 10 09:50:39 crc kubenswrapper[4715]: I1210 09:50:39.573539 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"05dd65e3-40ea-4d12-856f-3fa4f99712b2","Type":"ContainerDied","Data":"08e83a913dc6fc2c55e45154cc6ab39f9fd34e02a83bec96e5f874d4d85f996b"} Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.597981 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66bffaf0-4efe-435f-8092-a14c323cb97d","Type":"ContainerStarted","Data":"c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a"} Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.598408 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.601966 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"05dd65e3-40ea-4d12-856f-3fa4f99712b2","Type":"ContainerStarted","Data":"74050140b1df3a6e4a49961779ef3f1cffb694d0a93ed13ca56f0cd051eeccaa"} Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.605508 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-lz7jm" event={"ID":"6fc17771-af47-4ba4-be8a-6b3143a999f4","Type":"ContainerStarted","Data":"9c236e366b5bb8731b233f19a976e3865b22589359909f4421cd765ca833a28c"} Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.605593 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-lz7jm" Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.606936 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b4bfd970-df85-4d59-a211-ba4adbb501ee","Type":"ContainerStarted","Data":"040aa3e4ca4ce13ac29da91ff45a7bcb1e3d8f416bb2ad907496f976cf50fd9e"} Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.609330 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"219b8c44-8f39-40f1-9000-25404c35d495","Type":"ContainerStarted","Data":"caf5157b274e4b9219da989384d07ae9e150387878afd1310ebf18e172f88acf"} Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.611382 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rv2hq" event={"ID":"d6ae191b-a288-4185-b98b-15654b1e5dbf","Type":"ContainerStarted","Data":"d52a123156ba751fa5c4f22171461aa693effc7248429ba5ef6e2473762535f8"} Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.620611 4715 generic.go:334] "Generic (PLEG): container finished" podID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerID="88954537948f1b6b89f50c437a97d49e6c9e21fe9054281eefa526138a84ce0f" exitCode=0 Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.621386 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gclpm" event={"ID":"f65cc319-a237-48f7-a12a-96a923cfbea6","Type":"ContainerDied","Data":"88954537948f1b6b89f50c437a97d49e6c9e21fe9054281eefa526138a84ce0f"} Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.623899 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=21.949128706 podStartE2EDuration="28.623878469s" podCreationTimestamp="2025-12-10 09:50:14 +0000 UTC" firstStartedPulling="2025-12-10 09:50:34.894415001 +0000 UTC m=+997.637961252" lastFinishedPulling="2025-12-10 09:50:41.569164764 +0000 UTC m=+1004.312711015" observedRunningTime="2025-12-10 09:50:42.615128469 +0000 UTC m=+1005.358674730" watchObservedRunningTime="2025-12-10 09:50:42.623878469 +0000 UTC m=+1005.367424720" Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.627231 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8bcd31bd-4e40-41b6-92ef-6e0e0b170afa","Type":"ContainerStarted","Data":"8f13e01772e9459c989b29aa2a7321489314a1586e4f460f6bbc803c2d274eb1"} Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.664998 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-lz7jm" podStartSLOduration=18.331218015 podStartE2EDuration="24.664977446s" podCreationTimestamp="2025-12-10 09:50:18 +0000 UTC" firstStartedPulling="2025-12-10 09:50:35.250509377 +0000 UTC m=+997.994055628" lastFinishedPulling="2025-12-10 09:50:41.584268808 +0000 UTC m=+1004.327815059" observedRunningTime="2025-12-10 09:50:42.66364787 +0000 UTC m=+1005.407194121" watchObservedRunningTime="2025-12-10 09:50:42.664977446 +0000 UTC m=+1005.408523697" Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.698339 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=10.999933685 podStartE2EDuration="31.69831863s" podCreationTimestamp="2025-12-10 09:50:11 +0000 UTC" firstStartedPulling="2025-12-10 09:50:13.621933671 +0000 UTC m=+976.365479922" lastFinishedPulling="2025-12-10 09:50:34.320318616 +0000 UTC m=+997.063864867" observedRunningTime="2025-12-10 09:50:42.692130881 +0000 UTC m=+1005.435677132" watchObservedRunningTime="2025-12-10 09:50:42.69831863 +0000 UTC m=+1005.441864871" Dec 10 09:50:42 crc kubenswrapper[4715]: I1210 09:50:42.722266 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=10.505727767 podStartE2EDuration="32.722238177s" podCreationTimestamp="2025-12-10 09:50:10 +0000 UTC" firstStartedPulling="2025-12-10 09:50:12.0474375 +0000 UTC m=+974.790983751" lastFinishedPulling="2025-12-10 09:50:34.26394791 +0000 UTC m=+997.007494161" observedRunningTime="2025-12-10 09:50:42.71326527 +0000 UTC m=+1005.456811541" watchObservedRunningTime="2025-12-10 09:50:42.722238177 +0000 UTC m=+1005.465784428" Dec 10 09:50:43 crc kubenswrapper[4715]: I1210 09:50:43.010242 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:43 crc kubenswrapper[4715]: I1210 09:50:43.010297 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:43 crc kubenswrapper[4715]: I1210 09:50:43.337866 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 10 09:50:43 crc kubenswrapper[4715]: I1210 09:50:43.637807 4715 generic.go:334] "Generic (PLEG): container finished" podID="d6ae191b-a288-4185-b98b-15654b1e5dbf" containerID="d52a123156ba751fa5c4f22171461aa693effc7248429ba5ef6e2473762535f8" exitCode=0 Dec 10 09:50:43 crc kubenswrapper[4715]: I1210 09:50:43.638297 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rv2hq" event={"ID":"d6ae191b-a288-4185-b98b-15654b1e5dbf","Type":"ContainerDied","Data":"d52a123156ba751fa5c4f22171461aa693effc7248429ba5ef6e2473762535f8"} Dec 10 09:50:43 crc kubenswrapper[4715]: I1210 09:50:43.642063 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gclpm" event={"ID":"f65cc319-a237-48f7-a12a-96a923cfbea6","Type":"ContainerStarted","Data":"78202e248a04d770ebe3a4fd3d68441ed398792dba9edee3da4bd24ac8544bfb"} Dec 10 09:50:43 crc kubenswrapper[4715]: I1210 09:50:43.686536 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gclpm" podStartSLOduration=22.859512389 podStartE2EDuration="30.686516971s" podCreationTimestamp="2025-12-10 09:50:13 +0000 UTC" firstStartedPulling="2025-12-10 09:50:35.519394331 +0000 UTC m=+998.262940582" lastFinishedPulling="2025-12-10 09:50:43.346398903 +0000 UTC m=+1006.089945164" observedRunningTime="2025-12-10 09:50:43.682412068 +0000 UTC m=+1006.425958329" watchObservedRunningTime="2025-12-10 09:50:43.686516971 +0000 UTC m=+1006.430063222" Dec 10 09:50:44 crc kubenswrapper[4715]: I1210 09:50:44.199829 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:44 crc kubenswrapper[4715]: I1210 09:50:44.199878 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:44 crc kubenswrapper[4715]: I1210 09:50:44.522201 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:50:44 crc kubenswrapper[4715]: I1210 09:50:44.577924 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dj6bj"] Dec 10 09:50:44 crc kubenswrapper[4715]: I1210 09:50:44.650676 4715 generic.go:334] "Generic (PLEG): container finished" podID="ffb85f8e-eec3-4751-8ff2-31839610dba4" containerID="2e222994c8f64550c43d08feb9a1f5f6b036b48a0944cd422147ca51cacc8147" exitCode=0 Dec 10 09:50:44 crc kubenswrapper[4715]: I1210 09:50:44.650765 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" event={"ID":"ffb85f8e-eec3-4751-8ff2-31839610dba4","Type":"ContainerDied","Data":"2e222994c8f64550c43d08feb9a1f5f6b036b48a0944cd422147ca51cacc8147"} Dec 10 09:50:44 crc kubenswrapper[4715]: I1210 09:50:44.657731 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rv2hq" event={"ID":"d6ae191b-a288-4185-b98b-15654b1e5dbf","Type":"ContainerStarted","Data":"9aa6b547d9cd99ffa4e3578a011c219cd4dee4ca751324f57942c609370ea9c9"} Dec 10 09:50:44 crc kubenswrapper[4715]: I1210 09:50:44.657810 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rv2hq" event={"ID":"d6ae191b-a288-4185-b98b-15654b1e5dbf","Type":"ContainerStarted","Data":"916bf118092c5568d61dc1063f3c782b78cccde20663e6635170bbb293f20bf1"} Dec 10 09:50:44 crc kubenswrapper[4715]: I1210 09:50:44.700692 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-rv2hq" podStartSLOduration=21.1661075 podStartE2EDuration="26.700671624s" podCreationTimestamp="2025-12-10 09:50:18 +0000 UTC" firstStartedPulling="2025-12-10 09:50:36.049691494 +0000 UTC m=+998.793237735" lastFinishedPulling="2025-12-10 09:50:41.584255608 +0000 UTC m=+1004.327801859" observedRunningTime="2025-12-10 09:50:44.699767579 +0000 UTC m=+1007.443313830" watchObservedRunningTime="2025-12-10 09:50:44.700671624 +0000 UTC m=+1007.444217875" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.255950 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-gclpm" podUID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerName="registry-server" probeResult="failure" output=< Dec 10 09:50:45 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 10 09:50:45 crc kubenswrapper[4715]: > Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.504288 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-clmjq"] Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.505988 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.536762 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-clmjq"] Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.575068 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g27h4\" (UniqueName: \"kubernetes.io/projected/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-kube-api-access-g27h4\") pod \"dnsmasq-dns-7cb5889db5-clmjq\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.575134 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-config\") pod \"dnsmasq-dns-7cb5889db5-clmjq\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.575182 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-clmjq\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.666076 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.666201 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.677134 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g27h4\" (UniqueName: \"kubernetes.io/projected/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-kube-api-access-g27h4\") pod \"dnsmasq-dns-7cb5889db5-clmjq\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.677207 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-config\") pod \"dnsmasq-dns-7cb5889db5-clmjq\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.677257 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-clmjq\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.680753 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-config\") pod \"dnsmasq-dns-7cb5889db5-clmjq\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.685335 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-clmjq\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.710110 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g27h4\" (UniqueName: \"kubernetes.io/projected/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-kube-api-access-g27h4\") pod \"dnsmasq-dns-7cb5889db5-clmjq\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:45 crc kubenswrapper[4715]: E1210 09:50:45.795375 4715 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.201:34126->38.102.83.201:39121: read tcp 38.102.83.201:34126->38.102.83.201:39121: read: connection reset by peer Dec 10 09:50:45 crc kubenswrapper[4715]: I1210 09:50:45.843721 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.586308 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.623228 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.628548 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-6lcvb" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.628651 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.628904 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.629067 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.643582 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.750457 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/de6c4b14-17ae-4701-b7bc-b4d907f6964c-cache\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.750512 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.750563 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvxfx\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-kube-api-access-gvxfx\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.750647 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.750799 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/de6c4b14-17ae-4701-b7bc-b4d907f6964c-lock\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.852050 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/de6c4b14-17ae-4701-b7bc-b4d907f6964c-cache\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.852095 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.852112 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvxfx\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-kube-api-access-gvxfx\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.852129 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.852212 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/de6c4b14-17ae-4701-b7bc-b4d907f6964c-lock\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.852618 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/de6c4b14-17ae-4701-b7bc-b4d907f6964c-lock\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.852837 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/de6c4b14-17ae-4701-b7bc-b4d907f6964c-cache\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.853180 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: E1210 09:50:46.853427 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 09:50:46 crc kubenswrapper[4715]: E1210 09:50:46.853453 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 09:50:46 crc kubenswrapper[4715]: E1210 09:50:46.853498 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift podName:de6c4b14-17ae-4701-b7bc-b4d907f6964c nodeName:}" failed. No retries permitted until 2025-12-10 09:50:47.353482464 +0000 UTC m=+1010.097028715 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift") pod "swift-storage-0" (UID: "de6c4b14-17ae-4701-b7bc-b4d907f6964c") : configmap "swift-ring-files" not found Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.873866 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvxfx\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-kube-api-access-gvxfx\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:46 crc kubenswrapper[4715]: I1210 09:50:46.878925 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:47 crc kubenswrapper[4715]: I1210 09:50:47.365813 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:47 crc kubenswrapper[4715]: E1210 09:50:47.365991 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 09:50:47 crc kubenswrapper[4715]: E1210 09:50:47.366006 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 09:50:47 crc kubenswrapper[4715]: E1210 09:50:47.366059 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift podName:de6c4b14-17ae-4701-b7bc-b4d907f6964c nodeName:}" failed. No retries permitted until 2025-12-10 09:50:48.36604529 +0000 UTC m=+1011.109591541 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift") pod "swift-storage-0" (UID: "de6c4b14-17ae-4701-b7bc-b4d907f6964c") : configmap "swift-ring-files" not found Dec 10 09:50:47 crc kubenswrapper[4715]: I1210 09:50:47.714217 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:50:47 crc kubenswrapper[4715]: I1210 09:50:47.714279 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:50:47 crc kubenswrapper[4715]: I1210 09:50:47.714327 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:50:47 crc kubenswrapper[4715]: I1210 09:50:47.715036 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92c7512a624555c7386239654aadd7a7479838a440573c30b4fffcacf96a04bf"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 09:50:47 crc kubenswrapper[4715]: I1210 09:50:47.715108 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://92c7512a624555c7386239654aadd7a7479838a440573c30b4fffcacf96a04bf" gracePeriod=600 Dec 10 09:50:48 crc kubenswrapper[4715]: I1210 09:50:48.388496 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:48 crc kubenswrapper[4715]: E1210 09:50:48.388738 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 09:50:48 crc kubenswrapper[4715]: E1210 09:50:48.388954 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 09:50:48 crc kubenswrapper[4715]: E1210 09:50:48.389011 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift podName:de6c4b14-17ae-4701-b7bc-b4d907f6964c nodeName:}" failed. No retries permitted until 2025-12-10 09:50:50.388994745 +0000 UTC m=+1013.132540996 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift") pod "swift-storage-0" (UID: "de6c4b14-17ae-4701-b7bc-b4d907f6964c") : configmap "swift-ring-files" not found Dec 10 09:50:49 crc kubenswrapper[4715]: I1210 09:50:49.934784 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="92c7512a624555c7386239654aadd7a7479838a440573c30b4fffcacf96a04bf" exitCode=0 Dec 10 09:50:49 crc kubenswrapper[4715]: I1210 09:50:49.934838 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"92c7512a624555c7386239654aadd7a7479838a440573c30b4fffcacf96a04bf"} Dec 10 09:50:49 crc kubenswrapper[4715]: I1210 09:50:49.934890 4715 scope.go:117] "RemoveContainer" containerID="5d20ad65a2bc63bec106125489b213c4e46e5f275365a8160d4b290979ccc044" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.409700 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-47zsh"] Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.412822 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.416065 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.416256 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.417205 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.427126 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-47zsh"] Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.458263 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:50 crc kubenswrapper[4715]: E1210 09:50:50.458516 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 09:50:50 crc kubenswrapper[4715]: E1210 09:50:50.458535 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 09:50:50 crc kubenswrapper[4715]: E1210 09:50:50.458606 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift podName:de6c4b14-17ae-4701-b7bc-b4d907f6964c nodeName:}" failed. No retries permitted until 2025-12-10 09:50:54.458587472 +0000 UTC m=+1017.202133723 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift") pod "swift-storage-0" (UID: "de6c4b14-17ae-4701-b7bc-b4d907f6964c") : configmap "swift-ring-files" not found Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.559937 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-ring-data-devices\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.560024 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-dispersionconf\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.560126 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-swiftconf\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.560234 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-combined-ca-bundle\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.560323 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3a351985-c466-4a2b-b74e-f67d85fb5715-etc-swift\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.560415 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggrgq\" (UniqueName: \"kubernetes.io/projected/3a351985-c466-4a2b-b74e-f67d85fb5715-kube-api-access-ggrgq\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.560465 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-scripts\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.661551 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-ring-data-devices\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.661618 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-dispersionconf\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.661660 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-swiftconf\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.662449 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-ring-data-devices\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.661724 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-combined-ca-bundle\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.662649 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3a351985-c466-4a2b-b74e-f67d85fb5715-etc-swift\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.662709 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggrgq\" (UniqueName: \"kubernetes.io/projected/3a351985-c466-4a2b-b74e-f67d85fb5715-kube-api-access-ggrgq\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.662737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-scripts\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.663355 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-scripts\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.663646 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3a351985-c466-4a2b-b74e-f67d85fb5715-etc-swift\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.668821 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-combined-ca-bundle\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.671159 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-dispersionconf\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.671193 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-swiftconf\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.683775 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggrgq\" (UniqueName: \"kubernetes.io/projected/3a351985-c466-4a2b-b74e-f67d85fb5715-kube-api-access-ggrgq\") pod \"swift-ring-rebalance-47zsh\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:50 crc kubenswrapper[4715]: I1210 09:50:50.753325 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:50:51 crc kubenswrapper[4715]: I1210 09:50:51.466956 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 10 09:50:51 crc kubenswrapper[4715]: I1210 09:50:51.467343 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 10 09:50:53 crc kubenswrapper[4715]: I1210 09:50:53.036581 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 10 09:50:53 crc kubenswrapper[4715]: I1210 09:50:53.118429 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="8bcd31bd-4e40-41b6-92ef-6e0e0b170afa" containerName="galera" probeResult="failure" output=< Dec 10 09:50:53 crc kubenswrapper[4715]: wsrep_local_state_comment (Joined) differs from Synced Dec 10 09:50:53 crc kubenswrapper[4715]: > Dec 10 09:50:54 crc kubenswrapper[4715]: I1210 09:50:54.249511 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:54 crc kubenswrapper[4715]: I1210 09:50:54.295431 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:54 crc kubenswrapper[4715]: I1210 09:50:54.490631 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gclpm"] Dec 10 09:50:54 crc kubenswrapper[4715]: I1210 09:50:54.521798 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:50:54 crc kubenswrapper[4715]: E1210 09:50:54.521981 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 09:50:54 crc kubenswrapper[4715]: E1210 09:50:54.522011 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 09:50:54 crc kubenswrapper[4715]: E1210 09:50:54.522086 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift podName:de6c4b14-17ae-4701-b7bc-b4d907f6964c nodeName:}" failed. No retries permitted until 2025-12-10 09:51:02.522061362 +0000 UTC m=+1025.265607633 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift") pod "swift-storage-0" (UID: "de6c4b14-17ae-4701-b7bc-b4d907f6964c") : configmap "swift-ring-files" not found Dec 10 09:50:55 crc kubenswrapper[4715]: I1210 09:50:55.268613 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 10 09:50:55 crc kubenswrapper[4715]: I1210 09:50:55.977540 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gclpm" podUID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerName="registry-server" containerID="cri-o://78202e248a04d770ebe3a4fd3d68441ed398792dba9edee3da4bd24ac8544bfb" gracePeriod=2 Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.373972 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.412567 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-config\") pod \"ffb85f8e-eec3-4751-8ff2-31839610dba4\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.412720 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-dns-svc\") pod \"ffb85f8e-eec3-4751-8ff2-31839610dba4\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.412764 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpdtn\" (UniqueName: \"kubernetes.io/projected/ffb85f8e-eec3-4751-8ff2-31839610dba4-kube-api-access-kpdtn\") pod \"ffb85f8e-eec3-4751-8ff2-31839610dba4\" (UID: \"ffb85f8e-eec3-4751-8ff2-31839610dba4\") " Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.418457 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffb85f8e-eec3-4751-8ff2-31839610dba4-kube-api-access-kpdtn" (OuterVolumeSpecName: "kube-api-access-kpdtn") pod "ffb85f8e-eec3-4751-8ff2-31839610dba4" (UID: "ffb85f8e-eec3-4751-8ff2-31839610dba4"). InnerVolumeSpecName "kube-api-access-kpdtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.434366 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ffb85f8e-eec3-4751-8ff2-31839610dba4" (UID: "ffb85f8e-eec3-4751-8ff2-31839610dba4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.440471 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-config" (OuterVolumeSpecName: "config") pod "ffb85f8e-eec3-4751-8ff2-31839610dba4" (UID: "ffb85f8e-eec3-4751-8ff2-31839610dba4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.513988 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.514020 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffb85f8e-eec3-4751-8ff2-31839610dba4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.514029 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpdtn\" (UniqueName: \"kubernetes.io/projected/ffb85f8e-eec3-4751-8ff2-31839610dba4-kube-api-access-kpdtn\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.728997 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.859980 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.995894 4715 generic.go:334] "Generic (PLEG): container finished" podID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerID="78202e248a04d770ebe3a4fd3d68441ed398792dba9edee3da4bd24ac8544bfb" exitCode=0 Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.996197 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gclpm" event={"ID":"f65cc319-a237-48f7-a12a-96a923cfbea6","Type":"ContainerDied","Data":"78202e248a04d770ebe3a4fd3d68441ed398792dba9edee3da4bd24ac8544bfb"} Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.998220 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" event={"ID":"ffb85f8e-eec3-4751-8ff2-31839610dba4","Type":"ContainerDied","Data":"60e742c574034e39fbd4b89a33f45a8d6568fcb89a2b7cf82ecd47d143f5fa0f"} Dec 10 09:50:57 crc kubenswrapper[4715]: I1210 09:50:57.999347 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dj6bj" Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.047007 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dj6bj"] Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.056649 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dj6bj"] Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.075052 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-clmjq"] Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.377900 4715 scope.go:117] "RemoveContainer" containerID="2e222994c8f64550c43d08feb9a1f5f6b036b48a0944cd422147ca51cacc8147" Dec 10 09:50:58 crc kubenswrapper[4715]: W1210 09:50:58.388121 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee885b4e_7811_4e0c_9c7a_dde13f16b10b.slice/crio-a62cfb25c809dea6fe654b92b4f265695ac79de33b3b4796abe2b23a6f9f1a1c WatchSource:0}: Error finding container a62cfb25c809dea6fe654b92b4f265695ac79de33b3b4796abe2b23a6f9f1a1c: Status 404 returned error can't find the container with id a62cfb25c809dea6fe654b92b4f265695ac79de33b3b4796abe2b23a6f9f1a1c Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.594258 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.741492 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7mkc\" (UniqueName: \"kubernetes.io/projected/f65cc319-a237-48f7-a12a-96a923cfbea6-kube-api-access-g7mkc\") pod \"f65cc319-a237-48f7-a12a-96a923cfbea6\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.741759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-utilities\") pod \"f65cc319-a237-48f7-a12a-96a923cfbea6\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.741966 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-catalog-content\") pod \"f65cc319-a237-48f7-a12a-96a923cfbea6\" (UID: \"f65cc319-a237-48f7-a12a-96a923cfbea6\") " Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.743411 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-utilities" (OuterVolumeSpecName: "utilities") pod "f65cc319-a237-48f7-a12a-96a923cfbea6" (UID: "f65cc319-a237-48f7-a12a-96a923cfbea6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.746472 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f65cc319-a237-48f7-a12a-96a923cfbea6-kube-api-access-g7mkc" (OuterVolumeSpecName: "kube-api-access-g7mkc") pod "f65cc319-a237-48f7-a12a-96a923cfbea6" (UID: "f65cc319-a237-48f7-a12a-96a923cfbea6"). InnerVolumeSpecName "kube-api-access-g7mkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.760134 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f65cc319-a237-48f7-a12a-96a923cfbea6" (UID: "f65cc319-a237-48f7-a12a-96a923cfbea6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.798787 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-47zsh"] Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.844230 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7mkc\" (UniqueName: \"kubernetes.io/projected/f65cc319-a237-48f7-a12a-96a923cfbea6-kube-api-access-g7mkc\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.844270 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:58 crc kubenswrapper[4715]: I1210 09:50:58.844283 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f65cc319-a237-48f7-a12a-96a923cfbea6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:50:58 crc kubenswrapper[4715]: E1210 09:50:58.881697 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Dec 10 09:50:58 crc kubenswrapper[4715]: E1210 09:50:58.881875 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},EnvVar{Name:CONFIG_HASH,Value:n7hbhc9h66fh599h648h68ch55h678hch65ch655h66fh67fh68fh65dh669h58fh658h5dchfbh58dh697h657h675h696h684hch65bh5ch5bfhddq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mqsrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(219b8c44-8f39-40f1-9000-25404c35d495): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:50:58 crc kubenswrapper[4715]: E1210 09:50:58.883103 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="219b8c44-8f39-40f1-9000-25404c35d495" Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.008873 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-47zsh" event={"ID":"3a351985-c466-4a2b-b74e-f67d85fb5715","Type":"ContainerStarted","Data":"321f478ec960f5e41f07fb515a6eb67177732be7e9b10f3ff2e0ea60f67b89bc"} Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.010065 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" event={"ID":"ee885b4e-7811-4e0c-9c7a-dde13f16b10b","Type":"ContainerStarted","Data":"a62cfb25c809dea6fe654b92b4f265695ac79de33b3b4796abe2b23a6f9f1a1c"} Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.013372 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"13f540c3510be87c668f6f3667fb9d329e2ff72443e0f54c1a4f373c716a3669"} Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.016630 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gclpm" Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.016765 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gclpm" event={"ID":"f65cc319-a237-48f7-a12a-96a923cfbea6","Type":"ContainerDied","Data":"af2dc2732d9ad0b09cc98331b5540c5dbf3ca20c3e5835b1d0231e388edfb4a9"} Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.016808 4715 scope.go:117] "RemoveContainer" containerID="78202e248a04d770ebe3a4fd3d68441ed398792dba9edee3da4bd24ac8544bfb" Dec 10 09:50:59 crc kubenswrapper[4715]: E1210 09:50:59.017399 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="219b8c44-8f39-40f1-9000-25404c35d495" Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.035196 4715 scope.go:117] "RemoveContainer" containerID="88954537948f1b6b89f50c437a97d49e6c9e21fe9054281eefa526138a84ce0f" Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.056816 4715 scope.go:117] "RemoveContainer" containerID="4f0bfe2af2c33be5743d4e13bb9741870c9a9e12f4b24a331f20d4e415940f19" Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.077345 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gclpm"] Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.084503 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gclpm"] Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.688491 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f65cc319-a237-48f7-a12a-96a923cfbea6" path="/var/lib/kubelet/pods/f65cc319-a237-48f7-a12a-96a923cfbea6/volumes" Dec 10 09:50:59 crc kubenswrapper[4715]: I1210 09:50:59.690410 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffb85f8e-eec3-4751-8ff2-31839610dba4" path="/var/lib/kubelet/pods/ffb85f8e-eec3-4751-8ff2-31839610dba4/volumes" Dec 10 09:50:59 crc kubenswrapper[4715]: E1210 09:50:59.702091 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Dec 10 09:50:59 crc kubenswrapper[4715]: E1210 09:50:59.702361 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},EnvVar{Name:CONFIG_HASH,Value:n95h68ch57dhd9h64bh68dh79h5dh688h9bh57fh68bh685h5ffhfch5dh8fh666h78h59bhfchf6h5dh5d8h5f7h659h584h5b7h5cbh56h5bdhfq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vpt29,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(b4bfd970-df85-4d59-a211-ba4adbb501ee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:50:59 crc kubenswrapper[4715]: E1210 09:50:59.703540 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="b4bfd970-df85-4d59-a211-ba4adbb501ee" Dec 10 09:51:00 crc kubenswrapper[4715]: I1210 09:51:00.025978 4715 generic.go:334] "Generic (PLEG): container finished" podID="ee885b4e-7811-4e0c-9c7a-dde13f16b10b" containerID="8593ab29bf183ad7316ca98c595be21c61eafd1e2f17b90ba98f5a1b91d89b24" exitCode=0 Dec 10 09:51:00 crc kubenswrapper[4715]: I1210 09:51:00.026093 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" event={"ID":"ee885b4e-7811-4e0c-9c7a-dde13f16b10b","Type":"ContainerDied","Data":"8593ab29bf183ad7316ca98c595be21c61eafd1e2f17b90ba98f5a1b91d89b24"} Dec 10 09:51:00 crc kubenswrapper[4715]: E1210 09:51:00.029123 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="b4bfd970-df85-4d59-a211-ba4adbb501ee" Dec 10 09:51:01 crc kubenswrapper[4715]: I1210 09:51:01.058627 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" event={"ID":"ee885b4e-7811-4e0c-9c7a-dde13f16b10b","Type":"ContainerStarted","Data":"36297b25b4e40f5b783d6d93281da0514c5aa369e590e84b8971db5ef67bce72"} Dec 10 09:51:01 crc kubenswrapper[4715]: I1210 09:51:01.059182 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:51:01 crc kubenswrapper[4715]: I1210 09:51:01.084292 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" podStartSLOduration=16.084275739 podStartE2EDuration="16.084275739s" podCreationTimestamp="2025-12-10 09:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:01.073792001 +0000 UTC m=+1023.817338252" watchObservedRunningTime="2025-12-10 09:51:01.084275739 +0000 UTC m=+1023.827821990" Dec 10 09:51:01 crc kubenswrapper[4715]: I1210 09:51:01.420960 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 10 09:51:01 crc kubenswrapper[4715]: E1210 09:51:01.423124 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="b4bfd970-df85-4d59-a211-ba4adbb501ee" Dec 10 09:51:01 crc kubenswrapper[4715]: I1210 09:51:01.469670 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 10 09:51:01 crc kubenswrapper[4715]: I1210 09:51:01.546685 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 10 09:51:01 crc kubenswrapper[4715]: I1210 09:51:01.844491 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 10 09:51:01 crc kubenswrapper[4715]: E1210 09:51:01.846333 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="219b8c44-8f39-40f1-9000-25404c35d495" Dec 10 09:51:01 crc kubenswrapper[4715]: I1210 09:51:01.897372 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.071374 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.071534 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 10 09:51:02 crc kubenswrapper[4715]: E1210 09:51:02.072653 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="b4bfd970-df85-4d59-a211-ba4adbb501ee" Dec 10 09:51:02 crc kubenswrapper[4715]: E1210 09:51:02.073440 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="219b8c44-8f39-40f1-9000-25404c35d495" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.121609 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.124413 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.549161 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-p5m86"] Dec 10 09:51:02 crc kubenswrapper[4715]: E1210 09:51:02.549589 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffb85f8e-eec3-4751-8ff2-31839610dba4" containerName="init" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.549608 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffb85f8e-eec3-4751-8ff2-31839610dba4" containerName="init" Dec 10 09:51:02 crc kubenswrapper[4715]: E1210 09:51:02.549632 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerName="extract-utilities" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.549640 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerName="extract-utilities" Dec 10 09:51:02 crc kubenswrapper[4715]: E1210 09:51:02.549668 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerName="registry-server" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.549676 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerName="registry-server" Dec 10 09:51:02 crc kubenswrapper[4715]: E1210 09:51:02.549689 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerName="extract-content" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.549696 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerName="extract-content" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.549905 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f65cc319-a237-48f7-a12a-96a923cfbea6" containerName="registry-server" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.553105 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffb85f8e-eec3-4751-8ff2-31839610dba4" containerName="init" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.553861 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.556061 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.561163 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-p5m86"] Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.565126 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:51:02 crc kubenswrapper[4715]: E1210 09:51:02.565398 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 09:51:02 crc kubenswrapper[4715]: E1210 09:51:02.565416 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 09:51:02 crc kubenswrapper[4715]: E1210 09:51:02.565467 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift podName:de6c4b14-17ae-4701-b7bc-b4d907f6964c nodeName:}" failed. No retries permitted until 2025-12-10 09:51:18.565451129 +0000 UTC m=+1041.308997380 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift") pod "swift-storage-0" (UID: "de6c4b14-17ae-4701-b7bc-b4d907f6964c") : configmap "swift-ring-files" not found Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.667391 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-config\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.667446 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-ovs-rundir\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.667508 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.667580 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdnv9\" (UniqueName: \"kubernetes.io/projected/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-kube-api-access-qdnv9\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.667626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-combined-ca-bundle\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.667680 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-ovn-rundir\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.701514 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-clmjq"] Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.731152 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-dqcsf"] Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.732842 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.737771 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.755658 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-dqcsf"] Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.769465 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-ovn-rundir\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.769568 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-config\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.769600 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-ovs-rundir\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.769648 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.769677 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdnv9\" (UniqueName: \"kubernetes.io/projected/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-kube-api-access-qdnv9\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.769735 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-combined-ca-bundle\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.770405 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-ovs-rundir\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.770426 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-ovn-rundir\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.771099 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-config\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.775579 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.783495 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-combined-ca-bundle\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.795748 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdnv9\" (UniqueName: \"kubernetes.io/projected/e26b3d5b-1e63-4d22-80a5-55ffb58f664b-kube-api-access-qdnv9\") pod \"ovn-controller-metrics-p5m86\" (UID: \"e26b3d5b-1e63-4d22-80a5-55ffb58f664b\") " pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.873296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.873344 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-config\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.873438 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.873476 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kqtg\" (UniqueName: \"kubernetes.io/projected/bddce08b-71e0-44cb-9679-af43ffbef7c2-kube-api-access-4kqtg\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.881285 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-p5m86" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.975447 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kqtg\" (UniqueName: \"kubernetes.io/projected/bddce08b-71e0-44cb-9679-af43ffbef7c2-kube-api-access-4kqtg\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.975582 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.975609 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-config\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.975706 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.976477 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.980693 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-config\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.980759 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.993729 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f4e8-account-create-update-xbb2q"] Dec 10 09:51:02 crc kubenswrapper[4715]: I1210 09:51:02.995063 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f4e8-account-create-update-xbb2q" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.008849 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.010142 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-dqcsf"] Dec 10 09:51:03 crc kubenswrapper[4715]: E1210 09:51:03.010679 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-4kqtg], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" podUID="bddce08b-71e0-44cb-9679-af43ffbef7c2" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.017696 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kqtg\" (UniqueName: \"kubernetes.io/projected/bddce08b-71e0-44cb-9679-af43ffbef7c2-kube-api-access-4kqtg\") pod \"dnsmasq-dns-74f6f696b9-dqcsf\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.069996 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f4e8-account-create-update-xbb2q"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.076452 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-d462h"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.079360 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.080806 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h9sz\" (UniqueName: \"kubernetes.io/projected/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-kube-api-access-2h9sz\") pod \"keystone-f4e8-account-create-update-xbb2q\" (UID: \"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350\") " pod="openstack/keystone-f4e8-account-create-update-xbb2q" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.080862 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-operator-scripts\") pod \"keystone-f4e8-account-create-update-xbb2q\" (UID: \"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350\") " pod="openstack/keystone-f4e8-account-create-update-xbb2q" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.088404 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.090403 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" podUID="ee885b4e-7811-4e0c-9c7a-dde13f16b10b" containerName="dnsmasq-dns" containerID="cri-o://36297b25b4e40f5b783d6d93281da0514c5aa369e590e84b8971db5ef67bce72" gracePeriod=10 Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.091574 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:03 crc kubenswrapper[4715]: E1210 09:51:03.105563 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="219b8c44-8f39-40f1-9000-25404c35d495" Dec 10 09:51:03 crc kubenswrapper[4715]: E1210 09:51:03.105648 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="b4bfd970-df85-4d59-a211-ba4adbb501ee" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.126901 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-d462h"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.164988 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-lm8rz"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.166300 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lm8rz" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.183021 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lm8rz"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.184253 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-operator-scripts\") pod \"keystone-f4e8-account-create-update-xbb2q\" (UID: \"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350\") " pod="openstack/keystone-f4e8-account-create-update-xbb2q" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.184353 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.184381 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-config\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.184448 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-dns-svc\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.184511 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.184532 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwz4r\" (UniqueName: \"kubernetes.io/projected/de02f9b6-491d-4be4-9f06-7c1c53976694-kube-api-access-rwz4r\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.184586 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h9sz\" (UniqueName: \"kubernetes.io/projected/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-kube-api-access-2h9sz\") pod \"keystone-f4e8-account-create-update-xbb2q\" (UID: \"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350\") " pod="openstack/keystone-f4e8-account-create-update-xbb2q" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.185686 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-operator-scripts\") pod \"keystone-f4e8-account-create-update-xbb2q\" (UID: \"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350\") " pod="openstack/keystone-f4e8-account-create-update-xbb2q" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.214479 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.264878 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h9sz\" (UniqueName: \"kubernetes.io/projected/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-kube-api-access-2h9sz\") pod \"keystone-f4e8-account-create-update-xbb2q\" (UID: \"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350\") " pod="openstack/keystone-f4e8-account-create-update-xbb2q" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.286087 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-config\") pod \"bddce08b-71e0-44cb-9679-af43ffbef7c2\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.286386 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kqtg\" (UniqueName: \"kubernetes.io/projected/bddce08b-71e0-44cb-9679-af43ffbef7c2-kube-api-access-4kqtg\") pod \"bddce08b-71e0-44cb-9679-af43ffbef7c2\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.286527 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-dns-svc\") pod \"bddce08b-71e0-44cb-9679-af43ffbef7c2\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.286637 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-ovsdbserver-nb\") pod \"bddce08b-71e0-44cb-9679-af43ffbef7c2\" (UID: \"bddce08b-71e0-44cb-9679-af43ffbef7c2\") " Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.287531 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9291a4d5-3925-428d-9351-30e4bc4186c8-operator-scripts\") pod \"keystone-db-create-lm8rz\" (UID: \"9291a4d5-3925-428d-9351-30e4bc4186c8\") " pod="openstack/keystone-db-create-lm8rz" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.288836 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.288991 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57r9d\" (UniqueName: \"kubernetes.io/projected/9291a4d5-3925-428d-9351-30e4bc4186c8-kube-api-access-57r9d\") pod \"keystone-db-create-lm8rz\" (UID: \"9291a4d5-3925-428d-9351-30e4bc4186c8\") " pod="openstack/keystone-db-create-lm8rz" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.289087 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-config\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.289445 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-dns-svc\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.289570 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bddce08b-71e0-44cb-9679-af43ffbef7c2" (UID: "bddce08b-71e0-44cb-9679-af43ffbef7c2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.289635 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bddce08b-71e0-44cb-9679-af43ffbef7c2" (UID: "bddce08b-71e0-44cb-9679-af43ffbef7c2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.290892 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-dns-svc\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.291007 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.291162 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwz4r\" (UniqueName: \"kubernetes.io/projected/de02f9b6-491d-4be4-9f06-7c1c53976694-kube-api-access-rwz4r\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.291428 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.292673 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-config\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.292749 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.291777 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-ktxnt"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.292006 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-config" (OuterVolumeSpecName: "config") pod "bddce08b-71e0-44cb-9679-af43ffbef7c2" (UID: "bddce08b-71e0-44cb-9679-af43ffbef7c2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.290876 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.301588 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ktxnt" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.301901 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.302739 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ktxnt"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.323546 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bddce08b-71e0-44cb-9679-af43ffbef7c2-kube-api-access-4kqtg" (OuterVolumeSpecName: "kube-api-access-4kqtg") pod "bddce08b-71e0-44cb-9679-af43ffbef7c2" (UID: "bddce08b-71e0-44cb-9679-af43ffbef7c2"). InnerVolumeSpecName "kube-api-access-4kqtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.342796 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwz4r\" (UniqueName: \"kubernetes.io/projected/de02f9b6-491d-4be4-9f06-7c1c53976694-kube-api-access-rwz4r\") pod \"dnsmasq-dns-698758b865-d462h\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.380409 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6793-account-create-update-6f4cp"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.381659 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6793-account-create-update-6f4cp" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.384127 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.387562 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6793-account-create-update-6f4cp"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.396524 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4zvf\" (UniqueName: \"kubernetes.io/projected/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-kube-api-access-g4zvf\") pod \"placement-db-create-ktxnt\" (UID: \"d3e10b59-95b4-4a8c-b69d-392afceeb5cb\") " pod="openstack/placement-db-create-ktxnt" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.396598 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9291a4d5-3925-428d-9351-30e4bc4186c8-operator-scripts\") pod \"keystone-db-create-lm8rz\" (UID: \"9291a4d5-3925-428d-9351-30e4bc4186c8\") " pod="openstack/keystone-db-create-lm8rz" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.396705 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57r9d\" (UniqueName: \"kubernetes.io/projected/9291a4d5-3925-428d-9351-30e4bc4186c8-kube-api-access-57r9d\") pod \"keystone-db-create-lm8rz\" (UID: \"9291a4d5-3925-428d-9351-30e4bc4186c8\") " pod="openstack/keystone-db-create-lm8rz" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.396752 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-operator-scripts\") pod \"placement-db-create-ktxnt\" (UID: \"d3e10b59-95b4-4a8c-b69d-392afceeb5cb\") " pod="openstack/placement-db-create-ktxnt" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.396867 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bddce08b-71e0-44cb-9679-af43ffbef7c2-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.396881 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kqtg\" (UniqueName: \"kubernetes.io/projected/bddce08b-71e0-44cb-9679-af43ffbef7c2-kube-api-access-4kqtg\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.397594 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9291a4d5-3925-428d-9351-30e4bc4186c8-operator-scripts\") pod \"keystone-db-create-lm8rz\" (UID: \"9291a4d5-3925-428d-9351-30e4bc4186c8\") " pod="openstack/keystone-db-create-lm8rz" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.413072 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57r9d\" (UniqueName: \"kubernetes.io/projected/9291a4d5-3925-428d-9351-30e4bc4186c8-kube-api-access-57r9d\") pod \"keystone-db-create-lm8rz\" (UID: \"9291a4d5-3925-428d-9351-30e4bc4186c8\") " pod="openstack/keystone-db-create-lm8rz" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.445507 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f4e8-account-create-update-xbb2q" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.481062 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-7sk54"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.482184 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7sk54" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.493305 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-7sk54"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.498368 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7026670-8991-4a8f-b70d-3f3e6291fc6e-operator-scripts\") pod \"placement-6793-account-create-update-6f4cp\" (UID: \"f7026670-8991-4a8f-b70d-3f3e6291fc6e\") " pod="openstack/placement-6793-account-create-update-6f4cp" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.498436 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4zvf\" (UniqueName: \"kubernetes.io/projected/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-kube-api-access-g4zvf\") pod \"placement-db-create-ktxnt\" (UID: \"d3e10b59-95b4-4a8c-b69d-392afceeb5cb\") " pod="openstack/placement-db-create-ktxnt" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.498505 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-operator-scripts\") pod \"placement-db-create-ktxnt\" (UID: \"d3e10b59-95b4-4a8c-b69d-392afceeb5cb\") " pod="openstack/placement-db-create-ktxnt" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.498557 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgh8f\" (UniqueName: \"kubernetes.io/projected/f7026670-8991-4a8f-b70d-3f3e6291fc6e-kube-api-access-vgh8f\") pod \"placement-6793-account-create-update-6f4cp\" (UID: \"f7026670-8991-4a8f-b70d-3f3e6291fc6e\") " pod="openstack/placement-6793-account-create-update-6f4cp" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.499390 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-operator-scripts\") pod \"placement-db-create-ktxnt\" (UID: \"d3e10b59-95b4-4a8c-b69d-392afceeb5cb\") " pod="openstack/placement-db-create-ktxnt" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.507170 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.515798 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4zvf\" (UniqueName: \"kubernetes.io/projected/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-kube-api-access-g4zvf\") pod \"placement-db-create-ktxnt\" (UID: \"d3e10b59-95b4-4a8c-b69d-392afceeb5cb\") " pod="openstack/placement-db-create-ktxnt" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.539995 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lm8rz" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.571902 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6121-account-create-update-zw9vn"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.573708 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6121-account-create-update-zw9vn" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.576903 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.610896 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6gm9\" (UniqueName: \"kubernetes.io/projected/0d135565-25c1-467f-9556-c5114c704da0-kube-api-access-b6gm9\") pod \"glance-db-create-7sk54\" (UID: \"0d135565-25c1-467f-9556-c5114c704da0\") " pod="openstack/glance-db-create-7sk54" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.610960 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgh8f\" (UniqueName: \"kubernetes.io/projected/f7026670-8991-4a8f-b70d-3f3e6291fc6e-kube-api-access-vgh8f\") pod \"placement-6793-account-create-update-6f4cp\" (UID: \"f7026670-8991-4a8f-b70d-3f3e6291fc6e\") " pod="openstack/placement-6793-account-create-update-6f4cp" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.611010 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7026670-8991-4a8f-b70d-3f3e6291fc6e-operator-scripts\") pod \"placement-6793-account-create-update-6f4cp\" (UID: \"f7026670-8991-4a8f-b70d-3f3e6291fc6e\") " pod="openstack/placement-6793-account-create-update-6f4cp" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.611041 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d135565-25c1-467f-9556-c5114c704da0-operator-scripts\") pod \"glance-db-create-7sk54\" (UID: \"0d135565-25c1-467f-9556-c5114c704da0\") " pod="openstack/glance-db-create-7sk54" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.611941 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7026670-8991-4a8f-b70d-3f3e6291fc6e-operator-scripts\") pod \"placement-6793-account-create-update-6f4cp\" (UID: \"f7026670-8991-4a8f-b70d-3f3e6291fc6e\") " pod="openstack/placement-6793-account-create-update-6f4cp" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.628000 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgh8f\" (UniqueName: \"kubernetes.io/projected/f7026670-8991-4a8f-b70d-3f3e6291fc6e-kube-api-access-vgh8f\") pod \"placement-6793-account-create-update-6f4cp\" (UID: \"f7026670-8991-4a8f-b70d-3f3e6291fc6e\") " pod="openstack/placement-6793-account-create-update-6f4cp" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.633310 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6121-account-create-update-zw9vn"] Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.677629 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ktxnt" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.705818 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6793-account-create-update-6f4cp" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.712928 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6gm9\" (UniqueName: \"kubernetes.io/projected/0d135565-25c1-467f-9556-c5114c704da0-kube-api-access-b6gm9\") pod \"glance-db-create-7sk54\" (UID: \"0d135565-25c1-467f-9556-c5114c704da0\") " pod="openstack/glance-db-create-7sk54" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.713015 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9cbg\" (UniqueName: \"kubernetes.io/projected/44adefd5-a106-405e-b70c-464d3d111dfd-kube-api-access-g9cbg\") pod \"glance-6121-account-create-update-zw9vn\" (UID: \"44adefd5-a106-405e-b70c-464d3d111dfd\") " pod="openstack/glance-6121-account-create-update-zw9vn" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.713066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d135565-25c1-467f-9556-c5114c704da0-operator-scripts\") pod \"glance-db-create-7sk54\" (UID: \"0d135565-25c1-467f-9556-c5114c704da0\") " pod="openstack/glance-db-create-7sk54" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.713212 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44adefd5-a106-405e-b70c-464d3d111dfd-operator-scripts\") pod \"glance-6121-account-create-update-zw9vn\" (UID: \"44adefd5-a106-405e-b70c-464d3d111dfd\") " pod="openstack/glance-6121-account-create-update-zw9vn" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.714422 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d135565-25c1-467f-9556-c5114c704da0-operator-scripts\") pod \"glance-db-create-7sk54\" (UID: \"0d135565-25c1-467f-9556-c5114c704da0\") " pod="openstack/glance-db-create-7sk54" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.731820 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6gm9\" (UniqueName: \"kubernetes.io/projected/0d135565-25c1-467f-9556-c5114c704da0-kube-api-access-b6gm9\") pod \"glance-db-create-7sk54\" (UID: \"0d135565-25c1-467f-9556-c5114c704da0\") " pod="openstack/glance-db-create-7sk54" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.802747 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7sk54" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.814788 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44adefd5-a106-405e-b70c-464d3d111dfd-operator-scripts\") pod \"glance-6121-account-create-update-zw9vn\" (UID: \"44adefd5-a106-405e-b70c-464d3d111dfd\") " pod="openstack/glance-6121-account-create-update-zw9vn" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.814898 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9cbg\" (UniqueName: \"kubernetes.io/projected/44adefd5-a106-405e-b70c-464d3d111dfd-kube-api-access-g9cbg\") pod \"glance-6121-account-create-update-zw9vn\" (UID: \"44adefd5-a106-405e-b70c-464d3d111dfd\") " pod="openstack/glance-6121-account-create-update-zw9vn" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.815536 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44adefd5-a106-405e-b70c-464d3d111dfd-operator-scripts\") pod \"glance-6121-account-create-update-zw9vn\" (UID: \"44adefd5-a106-405e-b70c-464d3d111dfd\") " pod="openstack/glance-6121-account-create-update-zw9vn" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.854501 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9cbg\" (UniqueName: \"kubernetes.io/projected/44adefd5-a106-405e-b70c-464d3d111dfd-kube-api-access-g9cbg\") pod \"glance-6121-account-create-update-zw9vn\" (UID: \"44adefd5-a106-405e-b70c-464d3d111dfd\") " pod="openstack/glance-6121-account-create-update-zw9vn" Dec 10 09:51:03 crc kubenswrapper[4715]: I1210 09:51:03.910111 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6121-account-create-update-zw9vn" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.138057 4715 generic.go:334] "Generic (PLEG): container finished" podID="ee885b4e-7811-4e0c-9c7a-dde13f16b10b" containerID="36297b25b4e40f5b783d6d93281da0514c5aa369e590e84b8971db5ef67bce72" exitCode=0 Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.138284 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" event={"ID":"ee885b4e-7811-4e0c-9c7a-dde13f16b10b","Type":"ContainerDied","Data":"36297b25b4e40f5b783d6d93281da0514c5aa369e590e84b8971db5ef67bce72"} Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.138386 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-dqcsf" Dec 10 09:51:04 crc kubenswrapper[4715]: E1210 09:51:04.149892 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="219b8c44-8f39-40f1-9000-25404c35d495" Dec 10 09:51:04 crc kubenswrapper[4715]: E1210 09:51:04.167830 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="b4bfd970-df85-4d59-a211-ba4adbb501ee" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.192068 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jzv2h"] Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.197419 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.203949 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzv2h"] Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.313206 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.334354 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-utilities\") pod \"redhat-operators-jzv2h\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.334455 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-catalog-content\") pod \"redhat-operators-jzv2h\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.334572 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2jzf\" (UniqueName: \"kubernetes.io/projected/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-kube-api-access-r2jzf\") pod \"redhat-operators-jzv2h\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.357001 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-dqcsf"] Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.372010 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-dqcsf"] Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.387473 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-p5m86"] Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.436281 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-dns-svc\") pod \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.437222 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-config\") pod \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.437310 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g27h4\" (UniqueName: \"kubernetes.io/projected/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-kube-api-access-g27h4\") pod \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\" (UID: \"ee885b4e-7811-4e0c-9c7a-dde13f16b10b\") " Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.437683 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2jzf\" (UniqueName: \"kubernetes.io/projected/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-kube-api-access-r2jzf\") pod \"redhat-operators-jzv2h\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.437763 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-utilities\") pod \"redhat-operators-jzv2h\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.437859 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-catalog-content\") pod \"redhat-operators-jzv2h\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.438656 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-catalog-content\") pod \"redhat-operators-jzv2h\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.440549 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-utilities\") pod \"redhat-operators-jzv2h\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.443626 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-kube-api-access-g27h4" (OuterVolumeSpecName: "kube-api-access-g27h4") pod "ee885b4e-7811-4e0c-9c7a-dde13f16b10b" (UID: "ee885b4e-7811-4e0c-9c7a-dde13f16b10b"). InnerVolumeSpecName "kube-api-access-g27h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.472023 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2jzf\" (UniqueName: \"kubernetes.io/projected/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-kube-api-access-r2jzf\") pod \"redhat-operators-jzv2h\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.499282 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee885b4e-7811-4e0c-9c7a-dde13f16b10b" (UID: "ee885b4e-7811-4e0c-9c7a-dde13f16b10b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.532404 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-config" (OuterVolumeSpecName: "config") pod "ee885b4e-7811-4e0c-9c7a-dde13f16b10b" (UID: "ee885b4e-7811-4e0c-9c7a-dde13f16b10b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.539120 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.539150 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g27h4\" (UniqueName: \"kubernetes.io/projected/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-kube-api-access-g27h4\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.539161 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee885b4e-7811-4e0c-9c7a-dde13f16b10b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.553179 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lm8rz"] Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.568472 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-d462h"] Dec 10 09:51:04 crc kubenswrapper[4715]: W1210 09:51:04.574828 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde02f9b6_491d_4be4_9f06_7c1c53976694.slice/crio-fb8d87cfcd242f8575387dd7b6247e9f00d21c72d89c7fdb5ec1629c08d252d3 WatchSource:0}: Error finding container fb8d87cfcd242f8575387dd7b6247e9f00d21c72d89c7fdb5ec1629c08d252d3: Status 404 returned error can't find the container with id fb8d87cfcd242f8575387dd7b6247e9f00d21c72d89c7fdb5ec1629c08d252d3 Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.676450 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.763426 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f4e8-account-create-update-xbb2q"] Dec 10 09:51:04 crc kubenswrapper[4715]: W1210 09:51:04.772259 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd6bdd7f_e4fd_4c76_8f2e_286491ddb350.slice/crio-34ad014c67b955a04e04e48937019d2e058ac3f9cb4a365e838ca95de4e7e155 WatchSource:0}: Error finding container 34ad014c67b955a04e04e48937019d2e058ac3f9cb4a365e838ca95de4e7e155: Status 404 returned error can't find the container with id 34ad014c67b955a04e04e48937019d2e058ac3f9cb4a365e838ca95de4e7e155 Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.778665 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ktxnt"] Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.866543 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6793-account-create-update-6f4cp"] Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.899318 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-7sk54"] Dec 10 09:51:04 crc kubenswrapper[4715]: I1210 09:51:04.972845 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6121-account-create-update-zw9vn"] Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.204606 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzv2h"] Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.243150 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-p5m86" event={"ID":"e26b3d5b-1e63-4d22-80a5-55ffb58f664b","Type":"ContainerStarted","Data":"f5d558e664c6321b1d915bb8116beb87057d6e346a268e1eac64a7394259d757"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.251161 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f4e8-account-create-update-xbb2q" event={"ID":"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350","Type":"ContainerStarted","Data":"34ad014c67b955a04e04e48937019d2e058ac3f9cb4a365e838ca95de4e7e155"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.256935 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ktxnt" event={"ID":"d3e10b59-95b4-4a8c-b69d-392afceeb5cb","Type":"ContainerStarted","Data":"c6a85bc3031cf46e08695de5c7071ffdfd448e015a6ac3497793334fac618098"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.257759 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7sk54" event={"ID":"0d135565-25c1-467f-9556-c5114c704da0","Type":"ContainerStarted","Data":"d042dcfe195299436a81c3894504f94bc5ca8324725eb0f702fbd692ad454ed2"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.259523 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6121-account-create-update-zw9vn" event={"ID":"44adefd5-a106-405e-b70c-464d3d111dfd","Type":"ContainerStarted","Data":"52ad3a0fc7a3b1e58cddba63f2dd00430027dc44cb5a9bf6753cb7180c38f4d5"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.261898 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6793-account-create-update-6f4cp" event={"ID":"f7026670-8991-4a8f-b70d-3f3e6291fc6e","Type":"ContainerStarted","Data":"92885fd999fbfe0a53835785b5654b503d9fecd834a9573ead7d6888fd862b6a"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.270355 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-47zsh" event={"ID":"3a351985-c466-4a2b-b74e-f67d85fb5715","Type":"ContainerStarted","Data":"88eea8c7f5f8c15633aabada2952ae4f1a14062163d8115e0b42376f19107124"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.272304 4715 generic.go:334] "Generic (PLEG): container finished" podID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerID="bce68d6d483eb4de7af24eedf6792720982d1fdd5aa1661295c9e6ac5f66eaf9" exitCode=0 Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.272503 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-d462h" event={"ID":"de02f9b6-491d-4be4-9f06-7c1c53976694","Type":"ContainerDied","Data":"bce68d6d483eb4de7af24eedf6792720982d1fdd5aa1661295c9e6ac5f66eaf9"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.272533 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-d462h" event={"ID":"de02f9b6-491d-4be4-9f06-7c1c53976694","Type":"ContainerStarted","Data":"fb8d87cfcd242f8575387dd7b6247e9f00d21c72d89c7fdb5ec1629c08d252d3"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.274132 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lm8rz" event={"ID":"9291a4d5-3925-428d-9351-30e4bc4186c8","Type":"ContainerStarted","Data":"9134f9fb25c8bc74c9f5257182933ef4bff717392eda7190418d410117036b94"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.274158 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lm8rz" event={"ID":"9291a4d5-3925-428d-9351-30e4bc4186c8","Type":"ContainerStarted","Data":"2b83554240ed21c115f5ff1ef02bd418d44e5cd86d2857e7ec31744304e15e76"} Dec 10 09:51:05 crc kubenswrapper[4715]: W1210 09:51:05.284336 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01e46c2b_bbc9_46e8_bf17_ead30adb0f2c.slice/crio-ac92fff7bc1fdfb72b1e54b64f35022af9d4e197bd50b9586fb97002aa3b1f2a WatchSource:0}: Error finding container ac92fff7bc1fdfb72b1e54b64f35022af9d4e197bd50b9586fb97002aa3b1f2a: Status 404 returned error can't find the container with id ac92fff7bc1fdfb72b1e54b64f35022af9d4e197bd50b9586fb97002aa3b1f2a Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.284806 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" event={"ID":"ee885b4e-7811-4e0c-9c7a-dde13f16b10b","Type":"ContainerDied","Data":"a62cfb25c809dea6fe654b92b4f265695ac79de33b3b4796abe2b23a6f9f1a1c"} Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.284867 4715 scope.go:117] "RemoveContainer" containerID="36297b25b4e40f5b783d6d93281da0514c5aa369e590e84b8971db5ef67bce72" Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.285097 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-clmjq" Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.295369 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-47zsh" podStartSLOduration=10.356262963 podStartE2EDuration="15.295348575s" podCreationTimestamp="2025-12-10 09:50:50 +0000 UTC" firstStartedPulling="2025-12-10 09:50:58.799452808 +0000 UTC m=+1021.542999059" lastFinishedPulling="2025-12-10 09:51:03.73853841 +0000 UTC m=+1026.482084671" observedRunningTime="2025-12-10 09:51:05.293145334 +0000 UTC m=+1028.036691595" watchObservedRunningTime="2025-12-10 09:51:05.295348575 +0000 UTC m=+1028.038894846" Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.326650 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-lm8rz" podStartSLOduration=2.326627003 podStartE2EDuration="2.326627003s" podCreationTimestamp="2025-12-10 09:51:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:05.306593923 +0000 UTC m=+1028.050140174" watchObservedRunningTime="2025-12-10 09:51:05.326627003 +0000 UTC m=+1028.070173254" Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.414241 4715 scope.go:117] "RemoveContainer" containerID="8593ab29bf183ad7316ca98c595be21c61eafd1e2f17b90ba98f5a1b91d89b24" Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.421346 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-clmjq"] Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.428505 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-clmjq"] Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.631776 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bddce08b-71e0-44cb-9679-af43ffbef7c2" path="/var/lib/kubelet/pods/bddce08b-71e0-44cb-9679-af43ffbef7c2/volumes" Dec 10 09:51:05 crc kubenswrapper[4715]: I1210 09:51:05.632258 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee885b4e-7811-4e0c-9c7a-dde13f16b10b" path="/var/lib/kubelet/pods/ee885b4e-7811-4e0c-9c7a-dde13f16b10b/volumes" Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.292673 4715 generic.go:334] "Generic (PLEG): container finished" podID="bd6bdd7f-e4fd-4c76-8f2e-286491ddb350" containerID="ee91540b6671e2218402316ba4ab4386d6cb12134c441b70169d8512b75cd721" exitCode=0 Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.292724 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f4e8-account-create-update-xbb2q" event={"ID":"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350","Type":"ContainerDied","Data":"ee91540b6671e2218402316ba4ab4386d6cb12134c441b70169d8512b75cd721"} Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.294818 4715 generic.go:334] "Generic (PLEG): container finished" podID="9291a4d5-3925-428d-9351-30e4bc4186c8" containerID="9134f9fb25c8bc74c9f5257182933ef4bff717392eda7190418d410117036b94" exitCode=0 Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.294875 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lm8rz" event={"ID":"9291a4d5-3925-428d-9351-30e4bc4186c8","Type":"ContainerDied","Data":"9134f9fb25c8bc74c9f5257182933ef4bff717392eda7190418d410117036b94"} Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.296314 4715 generic.go:334] "Generic (PLEG): container finished" podID="0d135565-25c1-467f-9556-c5114c704da0" containerID="021b0cdf2963aa3d9ec5796b1f3bcd52205509f295c4c57361cd9d404a8693a5" exitCode=0 Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.296381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7sk54" event={"ID":"0d135565-25c1-467f-9556-c5114c704da0","Type":"ContainerDied","Data":"021b0cdf2963aa3d9ec5796b1f3bcd52205509f295c4c57361cd9d404a8693a5"} Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.297900 4715 generic.go:334] "Generic (PLEG): container finished" podID="44adefd5-a106-405e-b70c-464d3d111dfd" containerID="ec4511826d538604be5f4658736ea25f1e9ea24edcf9cad47ecd7104d46c4ad9" exitCode=0 Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.297967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6121-account-create-update-zw9vn" event={"ID":"44adefd5-a106-405e-b70c-464d3d111dfd","Type":"ContainerDied","Data":"ec4511826d538604be5f4658736ea25f1e9ea24edcf9cad47ecd7104d46c4ad9"} Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.299351 4715 generic.go:334] "Generic (PLEG): container finished" podID="d3e10b59-95b4-4a8c-b69d-392afceeb5cb" containerID="b4a38ed4e28f2aa3b65206788a62369dbc91ece4d8cc1f1f988e5c7347996839" exitCode=0 Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.299405 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ktxnt" event={"ID":"d3e10b59-95b4-4a8c-b69d-392afceeb5cb","Type":"ContainerDied","Data":"b4a38ed4e28f2aa3b65206788a62369dbc91ece4d8cc1f1f988e5c7347996839"} Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.300815 4715 generic.go:334] "Generic (PLEG): container finished" podID="f7026670-8991-4a8f-b70d-3f3e6291fc6e" containerID="0ff053fbe0e06f75d30846932e658aa9b03b2b4915164ce2d48cfca95161f99e" exitCode=0 Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.300871 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6793-account-create-update-6f4cp" event={"ID":"f7026670-8991-4a8f-b70d-3f3e6291fc6e","Type":"ContainerDied","Data":"0ff053fbe0e06f75d30846932e658aa9b03b2b4915164ce2d48cfca95161f99e"} Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.303837 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-p5m86" event={"ID":"e26b3d5b-1e63-4d22-80a5-55ffb58f664b","Type":"ContainerStarted","Data":"b6c90bf1593c5cdd636e71b89d865fe47ce9348ff7326fda4d1eca2091f06e22"} Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.306168 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-d462h" event={"ID":"de02f9b6-491d-4be4-9f06-7c1c53976694","Type":"ContainerStarted","Data":"82c73dd9e350a16ff73aac255bc07d2d8de274b87dacef3429c9d880d308c897"} Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.306770 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.309393 4715 generic.go:334] "Generic (PLEG): container finished" podID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerID="9a8b667d87a402e885f86907168ff41d5b7ab2b45e1c533664c49c5ae49ae013" exitCode=0 Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.309421 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzv2h" event={"ID":"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c","Type":"ContainerDied","Data":"9a8b667d87a402e885f86907168ff41d5b7ab2b45e1c533664c49c5ae49ae013"} Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.309485 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzv2h" event={"ID":"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c","Type":"ContainerStarted","Data":"ac92fff7bc1fdfb72b1e54b64f35022af9d4e197bd50b9586fb97002aa3b1f2a"} Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.434652 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-p5m86" podStartSLOduration=3.750354894 podStartE2EDuration="4.434627799s" podCreationTimestamp="2025-12-10 09:51:02 +0000 UTC" firstStartedPulling="2025-12-10 09:51:04.38975025 +0000 UTC m=+1027.133296501" lastFinishedPulling="2025-12-10 09:51:05.074023155 +0000 UTC m=+1027.817569406" observedRunningTime="2025-12-10 09:51:06.431225966 +0000 UTC m=+1029.174772217" watchObservedRunningTime="2025-12-10 09:51:06.434627799 +0000 UTC m=+1029.178174050" Dec 10 09:51:06 crc kubenswrapper[4715]: I1210 09:51:06.486636 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-d462h" podStartSLOduration=3.486617305 podStartE2EDuration="3.486617305s" podCreationTimestamp="2025-12-10 09:51:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:06.479424358 +0000 UTC m=+1029.222970609" watchObservedRunningTime="2025-12-10 09:51:06.486617305 +0000 UTC m=+1029.230163546" Dec 10 09:51:07 crc kubenswrapper[4715]: I1210 09:51:07.320210 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzv2h" event={"ID":"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c","Type":"ContainerStarted","Data":"0d66b62445cac79b9e76080cdfc5b4d9e3951ede5a1289448badd387bce34d52"} Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.208759 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6121-account-create-update-zw9vn" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.213379 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f4e8-account-create-update-xbb2q" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.223547 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lm8rz" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.227787 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ktxnt" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.249381 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6793-account-create-update-6f4cp" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.266786 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7sk54" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.270905 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7026670-8991-4a8f-b70d-3f3e6291fc6e-operator-scripts\") pod \"f7026670-8991-4a8f-b70d-3f3e6291fc6e\" (UID: \"f7026670-8991-4a8f-b70d-3f3e6291fc6e\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.270963 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9291a4d5-3925-428d-9351-30e4bc4186c8-operator-scripts\") pod \"9291a4d5-3925-428d-9351-30e4bc4186c8\" (UID: \"9291a4d5-3925-428d-9351-30e4bc4186c8\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.270993 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4zvf\" (UniqueName: \"kubernetes.io/projected/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-kube-api-access-g4zvf\") pod \"d3e10b59-95b4-4a8c-b69d-392afceeb5cb\" (UID: \"d3e10b59-95b4-4a8c-b69d-392afceeb5cb\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.271036 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-operator-scripts\") pod \"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350\" (UID: \"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.271096 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9cbg\" (UniqueName: \"kubernetes.io/projected/44adefd5-a106-405e-b70c-464d3d111dfd-kube-api-access-g9cbg\") pod \"44adefd5-a106-405e-b70c-464d3d111dfd\" (UID: \"44adefd5-a106-405e-b70c-464d3d111dfd\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.271119 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44adefd5-a106-405e-b70c-464d3d111dfd-operator-scripts\") pod \"44adefd5-a106-405e-b70c-464d3d111dfd\" (UID: \"44adefd5-a106-405e-b70c-464d3d111dfd\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.271192 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgh8f\" (UniqueName: \"kubernetes.io/projected/f7026670-8991-4a8f-b70d-3f3e6291fc6e-kube-api-access-vgh8f\") pod \"f7026670-8991-4a8f-b70d-3f3e6291fc6e\" (UID: \"f7026670-8991-4a8f-b70d-3f3e6291fc6e\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.271249 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57r9d\" (UniqueName: \"kubernetes.io/projected/9291a4d5-3925-428d-9351-30e4bc4186c8-kube-api-access-57r9d\") pod \"9291a4d5-3925-428d-9351-30e4bc4186c8\" (UID: \"9291a4d5-3925-428d-9351-30e4bc4186c8\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.271273 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h9sz\" (UniqueName: \"kubernetes.io/projected/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-kube-api-access-2h9sz\") pod \"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350\" (UID: \"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.271296 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-operator-scripts\") pod \"d3e10b59-95b4-4a8c-b69d-392afceeb5cb\" (UID: \"d3e10b59-95b4-4a8c-b69d-392afceeb5cb\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.273350 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3e10b59-95b4-4a8c-b69d-392afceeb5cb" (UID: "d3e10b59-95b4-4a8c-b69d-392afceeb5cb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.273742 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7026670-8991-4a8f-b70d-3f3e6291fc6e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7026670-8991-4a8f-b70d-3f3e6291fc6e" (UID: "f7026670-8991-4a8f-b70d-3f3e6291fc6e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.274128 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9291a4d5-3925-428d-9351-30e4bc4186c8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9291a4d5-3925-428d-9351-30e4bc4186c8" (UID: "9291a4d5-3925-428d-9351-30e4bc4186c8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.275861 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44adefd5-a106-405e-b70c-464d3d111dfd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "44adefd5-a106-405e-b70c-464d3d111dfd" (UID: "44adefd5-a106-405e-b70c-464d3d111dfd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.276551 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bd6bdd7f-e4fd-4c76-8f2e-286491ddb350" (UID: "bd6bdd7f-e4fd-4c76-8f2e-286491ddb350"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.292606 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-kube-api-access-2h9sz" (OuterVolumeSpecName: "kube-api-access-2h9sz") pod "bd6bdd7f-e4fd-4c76-8f2e-286491ddb350" (UID: "bd6bdd7f-e4fd-4c76-8f2e-286491ddb350"). InnerVolumeSpecName "kube-api-access-2h9sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.294121 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7026670-8991-4a8f-b70d-3f3e6291fc6e-kube-api-access-vgh8f" (OuterVolumeSpecName: "kube-api-access-vgh8f") pod "f7026670-8991-4a8f-b70d-3f3e6291fc6e" (UID: "f7026670-8991-4a8f-b70d-3f3e6291fc6e"). InnerVolumeSpecName "kube-api-access-vgh8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.295743 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9291a4d5-3925-428d-9351-30e4bc4186c8-kube-api-access-57r9d" (OuterVolumeSpecName: "kube-api-access-57r9d") pod "9291a4d5-3925-428d-9351-30e4bc4186c8" (UID: "9291a4d5-3925-428d-9351-30e4bc4186c8"). InnerVolumeSpecName "kube-api-access-57r9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.303072 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-kube-api-access-g4zvf" (OuterVolumeSpecName: "kube-api-access-g4zvf") pod "d3e10b59-95b4-4a8c-b69d-392afceeb5cb" (UID: "d3e10b59-95b4-4a8c-b69d-392afceeb5cb"). InnerVolumeSpecName "kube-api-access-g4zvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.317320 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44adefd5-a106-405e-b70c-464d3d111dfd-kube-api-access-g9cbg" (OuterVolumeSpecName: "kube-api-access-g9cbg") pod "44adefd5-a106-405e-b70c-464d3d111dfd" (UID: "44adefd5-a106-405e-b70c-464d3d111dfd"). InnerVolumeSpecName "kube-api-access-g9cbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.334068 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6121-account-create-update-zw9vn" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.335092 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6121-account-create-update-zw9vn" event={"ID":"44adefd5-a106-405e-b70c-464d3d111dfd","Type":"ContainerDied","Data":"52ad3a0fc7a3b1e58cddba63f2dd00430027dc44cb5a9bf6753cb7180c38f4d5"} Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.335167 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52ad3a0fc7a3b1e58cddba63f2dd00430027dc44cb5a9bf6753cb7180c38f4d5" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.339447 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ktxnt" event={"ID":"d3e10b59-95b4-4a8c-b69d-392afceeb5cb","Type":"ContainerDied","Data":"c6a85bc3031cf46e08695de5c7071ffdfd448e015a6ac3497793334fac618098"} Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.339484 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6a85bc3031cf46e08695de5c7071ffdfd448e015a6ac3497793334fac618098" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.339481 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ktxnt" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.341289 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6793-account-create-update-6f4cp" event={"ID":"f7026670-8991-4a8f-b70d-3f3e6291fc6e","Type":"ContainerDied","Data":"92885fd999fbfe0a53835785b5654b503d9fecd834a9573ead7d6888fd862b6a"} Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.341406 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92885fd999fbfe0a53835785b5654b503d9fecd834a9573ead7d6888fd862b6a" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.341518 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6793-account-create-update-6f4cp" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.346254 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f4e8-account-create-update-xbb2q" event={"ID":"bd6bdd7f-e4fd-4c76-8f2e-286491ddb350","Type":"ContainerDied","Data":"34ad014c67b955a04e04e48937019d2e058ac3f9cb4a365e838ca95de4e7e155"} Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.346301 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34ad014c67b955a04e04e48937019d2e058ac3f9cb4a365e838ca95de4e7e155" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.346414 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f4e8-account-create-update-xbb2q" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.363374 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lm8rz" event={"ID":"9291a4d5-3925-428d-9351-30e4bc4186c8","Type":"ContainerDied","Data":"2b83554240ed21c115f5ff1ef02bd418d44e5cd86d2857e7ec31744304e15e76"} Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.363415 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lm8rz" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.363421 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b83554240ed21c115f5ff1ef02bd418d44e5cd86d2857e7ec31744304e15e76" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.365518 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7sk54" event={"ID":"0d135565-25c1-467f-9556-c5114c704da0","Type":"ContainerDied","Data":"d042dcfe195299436a81c3894504f94bc5ca8324725eb0f702fbd692ad454ed2"} Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.365541 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d042dcfe195299436a81c3894504f94bc5ca8324725eb0f702fbd692ad454ed2" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.365578 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7sk54" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.374347 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d135565-25c1-467f-9556-c5114c704da0-operator-scripts\") pod \"0d135565-25c1-467f-9556-c5114c704da0\" (UID: \"0d135565-25c1-467f-9556-c5114c704da0\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.374455 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6gm9\" (UniqueName: \"kubernetes.io/projected/0d135565-25c1-467f-9556-c5114c704da0-kube-api-access-b6gm9\") pod \"0d135565-25c1-467f-9556-c5114c704da0\" (UID: \"0d135565-25c1-467f-9556-c5114c704da0\") " Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.374764 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d135565-25c1-467f-9556-c5114c704da0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d135565-25c1-467f-9556-c5114c704da0" (UID: "0d135565-25c1-467f-9556-c5114c704da0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375065 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h9sz\" (UniqueName: \"kubernetes.io/projected/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-kube-api-access-2h9sz\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375092 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375108 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7026670-8991-4a8f-b70d-3f3e6291fc6e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375122 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9291a4d5-3925-428d-9351-30e4bc4186c8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375139 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4zvf\" (UniqueName: \"kubernetes.io/projected/d3e10b59-95b4-4a8c-b69d-392afceeb5cb-kube-api-access-g4zvf\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375153 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d135565-25c1-467f-9556-c5114c704da0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375165 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375178 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9cbg\" (UniqueName: \"kubernetes.io/projected/44adefd5-a106-405e-b70c-464d3d111dfd-kube-api-access-g9cbg\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375190 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44adefd5-a106-405e-b70c-464d3d111dfd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375202 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgh8f\" (UniqueName: \"kubernetes.io/projected/f7026670-8991-4a8f-b70d-3f3e6291fc6e-kube-api-access-vgh8f\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.375213 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57r9d\" (UniqueName: \"kubernetes.io/projected/9291a4d5-3925-428d-9351-30e4bc4186c8-kube-api-access-57r9d\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.387574 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d135565-25c1-467f-9556-c5114c704da0-kube-api-access-b6gm9" (OuterVolumeSpecName: "kube-api-access-b6gm9") pod "0d135565-25c1-467f-9556-c5114c704da0" (UID: "0d135565-25c1-467f-9556-c5114c704da0"). InnerVolumeSpecName "kube-api-access-b6gm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:08 crc kubenswrapper[4715]: I1210 09:51:08.476639 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6gm9\" (UniqueName: \"kubernetes.io/projected/0d135565-25c1-467f-9556-c5114c704da0-kube-api-access-b6gm9\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:11 crc kubenswrapper[4715]: I1210 09:51:11.388267 4715 generic.go:334] "Generic (PLEG): container finished" podID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerID="0d66b62445cac79b9e76080cdfc5b4d9e3951ede5a1289448badd387bce34d52" exitCode=0 Dec 10 09:51:11 crc kubenswrapper[4715]: I1210 09:51:11.388454 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzv2h" event={"ID":"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c","Type":"ContainerDied","Data":"0d66b62445cac79b9e76080cdfc5b4d9e3951ede5a1289448badd387bce34d52"} Dec 10 09:51:11 crc kubenswrapper[4715]: I1210 09:51:11.391187 4715 generic.go:334] "Generic (PLEG): container finished" podID="5afc401a-b98d-4b26-8a47-c18582062f49" containerID="993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c" exitCode=0 Dec 10 09:51:11 crc kubenswrapper[4715]: I1210 09:51:11.391207 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5afc401a-b98d-4b26-8a47-c18582062f49","Type":"ContainerDied","Data":"993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c"} Dec 10 09:51:12 crc kubenswrapper[4715]: I1210 09:51:12.407901 4715 generic.go:334] "Generic (PLEG): container finished" podID="ad27e35f-aceb-42a8-a60f-efd71d0a3b73" containerID="ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2" exitCode=0 Dec 10 09:51:12 crc kubenswrapper[4715]: I1210 09:51:12.407942 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad27e35f-aceb-42a8-a60f-efd71d0a3b73","Type":"ContainerDied","Data":"ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2"} Dec 10 09:51:12 crc kubenswrapper[4715]: I1210 09:51:12.415949 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5afc401a-b98d-4b26-8a47-c18582062f49","Type":"ContainerStarted","Data":"e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b"} Dec 10 09:51:12 crc kubenswrapper[4715]: I1210 09:51:12.416347 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:51:12 crc kubenswrapper[4715]: I1210 09:51:12.476526 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.530460467 podStartE2EDuration="1m3.476502485s" podCreationTimestamp="2025-12-10 09:50:09 +0000 UTC" firstStartedPulling="2025-12-10 09:50:11.325030268 +0000 UTC m=+974.068576519" lastFinishedPulling="2025-12-10 09:50:34.271072286 +0000 UTC m=+997.014618537" observedRunningTime="2025-12-10 09:51:12.464172927 +0000 UTC m=+1035.207719188" watchObservedRunningTime="2025-12-10 09:51:12.476502485 +0000 UTC m=+1035.220048736" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.434051 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzv2h" event={"ID":"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c","Type":"ContainerStarted","Data":"d4b28f8243a1f77d4aecaad28451bb4efb5696b8f113462baccda860b55b1340"} Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.436791 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad27e35f-aceb-42a8-a60f-efd71d0a3b73","Type":"ContainerStarted","Data":"64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653"} Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.465606 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jzv2h" podStartSLOduration=3.428475211 podStartE2EDuration="9.465578046s" podCreationTimestamp="2025-12-10 09:51:04 +0000 UTC" firstStartedPulling="2025-12-10 09:51:06.311286757 +0000 UTC m=+1029.054833008" lastFinishedPulling="2025-12-10 09:51:12.348389592 +0000 UTC m=+1035.091935843" observedRunningTime="2025-12-10 09:51:13.451218449 +0000 UTC m=+1036.194764700" watchObservedRunningTime="2025-12-10 09:51:13.465578046 +0000 UTC m=+1036.209124337" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.479293 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=42.259580721 podStartE2EDuration="1m5.479275473s" podCreationTimestamp="2025-12-10 09:50:08 +0000 UTC" firstStartedPulling="2025-12-10 09:50:11.10666293 +0000 UTC m=+973.850209181" lastFinishedPulling="2025-12-10 09:50:34.326357672 +0000 UTC m=+997.069903933" observedRunningTime="2025-12-10 09:51:13.470290719 +0000 UTC m=+1036.213836970" watchObservedRunningTime="2025-12-10 09:51:13.479275473 +0000 UTC m=+1036.222821724" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.508160 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.578492 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lm5b9"] Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.578978 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" podUID="9fbc1385-27ee-4c8b-8ce5-de03f14b4876" containerName="dnsmasq-dns" containerID="cri-o://fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692" gracePeriod=10 Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.875877 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-gd24z"] Dec 10 09:51:13 crc kubenswrapper[4715]: E1210 09:51:13.876497 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee885b4e-7811-4e0c-9c7a-dde13f16b10b" containerName="init" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876513 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee885b4e-7811-4e0c-9c7a-dde13f16b10b" containerName="init" Dec 10 09:51:13 crc kubenswrapper[4715]: E1210 09:51:13.876527 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee885b4e-7811-4e0c-9c7a-dde13f16b10b" containerName="dnsmasq-dns" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876533 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee885b4e-7811-4e0c-9c7a-dde13f16b10b" containerName="dnsmasq-dns" Dec 10 09:51:13 crc kubenswrapper[4715]: E1210 09:51:13.876547 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7026670-8991-4a8f-b70d-3f3e6291fc6e" containerName="mariadb-account-create-update" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876554 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7026670-8991-4a8f-b70d-3f3e6291fc6e" containerName="mariadb-account-create-update" Dec 10 09:51:13 crc kubenswrapper[4715]: E1210 09:51:13.876564 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44adefd5-a106-405e-b70c-464d3d111dfd" containerName="mariadb-account-create-update" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876571 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="44adefd5-a106-405e-b70c-464d3d111dfd" containerName="mariadb-account-create-update" Dec 10 09:51:13 crc kubenswrapper[4715]: E1210 09:51:13.876580 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d135565-25c1-467f-9556-c5114c704da0" containerName="mariadb-database-create" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876585 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d135565-25c1-467f-9556-c5114c704da0" containerName="mariadb-database-create" Dec 10 09:51:13 crc kubenswrapper[4715]: E1210 09:51:13.876596 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9291a4d5-3925-428d-9351-30e4bc4186c8" containerName="mariadb-database-create" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876602 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9291a4d5-3925-428d-9351-30e4bc4186c8" containerName="mariadb-database-create" Dec 10 09:51:13 crc kubenswrapper[4715]: E1210 09:51:13.876619 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e10b59-95b4-4a8c-b69d-392afceeb5cb" containerName="mariadb-database-create" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876626 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e10b59-95b4-4a8c-b69d-392afceeb5cb" containerName="mariadb-database-create" Dec 10 09:51:13 crc kubenswrapper[4715]: E1210 09:51:13.876633 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd6bdd7f-e4fd-4c76-8f2e-286491ddb350" containerName="mariadb-account-create-update" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876639 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd6bdd7f-e4fd-4c76-8f2e-286491ddb350" containerName="mariadb-account-create-update" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876796 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e10b59-95b4-4a8c-b69d-392afceeb5cb" containerName="mariadb-database-create" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876808 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d135565-25c1-467f-9556-c5114c704da0" containerName="mariadb-database-create" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876817 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd6bdd7f-e4fd-4c76-8f2e-286491ddb350" containerName="mariadb-account-create-update" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876827 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee885b4e-7811-4e0c-9c7a-dde13f16b10b" containerName="dnsmasq-dns" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876839 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="44adefd5-a106-405e-b70c-464d3d111dfd" containerName="mariadb-account-create-update" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876849 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9291a4d5-3925-428d-9351-30e4bc4186c8" containerName="mariadb-database-create" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.876855 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7026670-8991-4a8f-b70d-3f3e6291fc6e" containerName="mariadb-account-create-update" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.877431 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.885477 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.891397 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gckbz" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.965790 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-gd24z"] Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.969692 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-config-data\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.969787 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-combined-ca-bundle\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.969843 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2shzm\" (UniqueName: \"kubernetes.io/projected/8ad144d0-9a73-4736-8f88-43dd295478d7-kube-api-access-2shzm\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:13 crc kubenswrapper[4715]: I1210 09:51:13.969875 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-db-sync-config-data\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.028147 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-lz7jm" podUID="6fc17771-af47-4ba4-be8a-6b3143a999f4" containerName="ovn-controller" probeResult="failure" output=< Dec 10 09:51:14 crc kubenswrapper[4715]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 10 09:51:14 crc kubenswrapper[4715]: > Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.074904 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-combined-ca-bundle\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.075031 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2shzm\" (UniqueName: \"kubernetes.io/projected/8ad144d0-9a73-4736-8f88-43dd295478d7-kube-api-access-2shzm\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.075074 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-db-sync-config-data\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.075121 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-config-data\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.083017 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-combined-ca-bundle\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.104816 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-config-data\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.111910 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2shzm\" (UniqueName: \"kubernetes.io/projected/8ad144d0-9a73-4736-8f88-43dd295478d7-kube-api-access-2shzm\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.115501 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-db-sync-config-data\") pod \"glance-db-sync-gd24z\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.241782 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.275007 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.389799 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwh7h\" (UniqueName: \"kubernetes.io/projected/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-kube-api-access-mwh7h\") pod \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.390264 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-config\") pod \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.390324 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-dns-svc\") pod \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\" (UID: \"9fbc1385-27ee-4c8b-8ce5-de03f14b4876\") " Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.403146 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-kube-api-access-mwh7h" (OuterVolumeSpecName: "kube-api-access-mwh7h") pod "9fbc1385-27ee-4c8b-8ce5-de03f14b4876" (UID: "9fbc1385-27ee-4c8b-8ce5-de03f14b4876"). InnerVolumeSpecName "kube-api-access-mwh7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.447515 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-config" (OuterVolumeSpecName: "config") pod "9fbc1385-27ee-4c8b-8ce5-de03f14b4876" (UID: "9fbc1385-27ee-4c8b-8ce5-de03f14b4876"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.453484 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9fbc1385-27ee-4c8b-8ce5-de03f14b4876" (UID: "9fbc1385-27ee-4c8b-8ce5-de03f14b4876"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.476281 4715 generic.go:334] "Generic (PLEG): container finished" podID="9fbc1385-27ee-4c8b-8ce5-de03f14b4876" containerID="fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692" exitCode=0 Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.476333 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" event={"ID":"9fbc1385-27ee-4c8b-8ce5-de03f14b4876","Type":"ContainerDied","Data":"fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692"} Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.476371 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" event={"ID":"9fbc1385-27ee-4c8b-8ce5-de03f14b4876","Type":"ContainerDied","Data":"7e96b664c205048e259ec4b4af9d546a412e800373f0c95be102623cc23685b2"} Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.476394 4715 scope.go:117] "RemoveContainer" containerID="fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.476419 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lm5b9" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.494144 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.494169 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.494181 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwh7h\" (UniqueName: \"kubernetes.io/projected/9fbc1385-27ee-4c8b-8ce5-de03f14b4876-kube-api-access-mwh7h\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.509360 4715 scope.go:117] "RemoveContainer" containerID="5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.517229 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lm5b9"] Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.531191 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lm5b9"] Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.536383 4715 scope.go:117] "RemoveContainer" containerID="fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692" Dec 10 09:51:14 crc kubenswrapper[4715]: E1210 09:51:14.538521 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692\": container with ID starting with fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692 not found: ID does not exist" containerID="fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.538592 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692"} err="failed to get container status \"fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692\": rpc error: code = NotFound desc = could not find container \"fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692\": container with ID starting with fcd14e1ce2369de47e06cde7eb910478ab4cb8172116cad8d07eb25108256692 not found: ID does not exist" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.538625 4715 scope.go:117] "RemoveContainer" containerID="5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5" Dec 10 09:51:14 crc kubenswrapper[4715]: E1210 09:51:14.540223 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5\": container with ID starting with 5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5 not found: ID does not exist" containerID="5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.540278 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5"} err="failed to get container status \"5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5\": rpc error: code = NotFound desc = could not find container \"5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5\": container with ID starting with 5a792a84e415b81f8efe9bd98bb50a0c73b5010ea674d188a33a29b2349d9df5 not found: ID does not exist" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.677107 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.677330 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:14 crc kubenswrapper[4715]: I1210 09:51:14.959107 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-gd24z"] Dec 10 09:51:15 crc kubenswrapper[4715]: I1210 09:51:15.487017 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gd24z" event={"ID":"8ad144d0-9a73-4736-8f88-43dd295478d7","Type":"ContainerStarted","Data":"1d722e9e1d0bddadb9315d761022d958586c6bc94155e1ba086a5615c1285381"} Dec 10 09:51:15 crc kubenswrapper[4715]: I1210 09:51:15.613278 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fbc1385-27ee-4c8b-8ce5-de03f14b4876" path="/var/lib/kubelet/pods/9fbc1385-27ee-4c8b-8ce5-de03f14b4876/volumes" Dec 10 09:51:15 crc kubenswrapper[4715]: I1210 09:51:15.725396 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jzv2h" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerName="registry-server" probeResult="failure" output=< Dec 10 09:51:15 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 10 09:51:15 crc kubenswrapper[4715]: > Dec 10 09:51:16 crc kubenswrapper[4715]: I1210 09:51:16.497104 4715 generic.go:334] "Generic (PLEG): container finished" podID="3a351985-c466-4a2b-b74e-f67d85fb5715" containerID="88eea8c7f5f8c15633aabada2952ae4f1a14062163d8115e0b42376f19107124" exitCode=0 Dec 10 09:51:16 crc kubenswrapper[4715]: I1210 09:51:16.497190 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-47zsh" event={"ID":"3a351985-c466-4a2b-b74e-f67d85fb5715","Type":"ContainerDied","Data":"88eea8c7f5f8c15633aabada2952ae4f1a14062163d8115e0b42376f19107124"} Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.850091 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.961381 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3a351985-c466-4a2b-b74e-f67d85fb5715-etc-swift\") pod \"3a351985-c466-4a2b-b74e-f67d85fb5715\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.961477 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggrgq\" (UniqueName: \"kubernetes.io/projected/3a351985-c466-4a2b-b74e-f67d85fb5715-kube-api-access-ggrgq\") pod \"3a351985-c466-4a2b-b74e-f67d85fb5715\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.961509 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-dispersionconf\") pod \"3a351985-c466-4a2b-b74e-f67d85fb5715\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.961606 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-scripts\") pod \"3a351985-c466-4a2b-b74e-f67d85fb5715\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.961626 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-ring-data-devices\") pod \"3a351985-c466-4a2b-b74e-f67d85fb5715\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.961724 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-combined-ca-bundle\") pod \"3a351985-c466-4a2b-b74e-f67d85fb5715\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.961745 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-swiftconf\") pod \"3a351985-c466-4a2b-b74e-f67d85fb5715\" (UID: \"3a351985-c466-4a2b-b74e-f67d85fb5715\") " Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.963593 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a351985-c466-4a2b-b74e-f67d85fb5715-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3a351985-c466-4a2b-b74e-f67d85fb5715" (UID: "3a351985-c466-4a2b-b74e-f67d85fb5715"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.963888 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3a351985-c466-4a2b-b74e-f67d85fb5715" (UID: "3a351985-c466-4a2b-b74e-f67d85fb5715"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.968619 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a351985-c466-4a2b-b74e-f67d85fb5715-kube-api-access-ggrgq" (OuterVolumeSpecName: "kube-api-access-ggrgq") pod "3a351985-c466-4a2b-b74e-f67d85fb5715" (UID: "3a351985-c466-4a2b-b74e-f67d85fb5715"). InnerVolumeSpecName "kube-api-access-ggrgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.986681 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3a351985-c466-4a2b-b74e-f67d85fb5715" (UID: "3a351985-c466-4a2b-b74e-f67d85fb5715"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:17 crc kubenswrapper[4715]: I1210 09:51:17.989319 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3a351985-c466-4a2b-b74e-f67d85fb5715" (UID: "3a351985-c466-4a2b-b74e-f67d85fb5715"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.007226 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a351985-c466-4a2b-b74e-f67d85fb5715" (UID: "3a351985-c466-4a2b-b74e-f67d85fb5715"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.012296 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-scripts" (OuterVolumeSpecName: "scripts") pod "3a351985-c466-4a2b-b74e-f67d85fb5715" (UID: "3a351985-c466-4a2b-b74e-f67d85fb5715"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.064117 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.064168 4715 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.064180 4715 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3a351985-c466-4a2b-b74e-f67d85fb5715-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.064194 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggrgq\" (UniqueName: \"kubernetes.io/projected/3a351985-c466-4a2b-b74e-f67d85fb5715-kube-api-access-ggrgq\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.064207 4715 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3a351985-c466-4a2b-b74e-f67d85fb5715-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.064219 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.064231 4715 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3a351985-c466-4a2b-b74e-f67d85fb5715-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.515114 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-47zsh" event={"ID":"3a351985-c466-4a2b-b74e-f67d85fb5715","Type":"ContainerDied","Data":"321f478ec960f5e41f07fb515a6eb67177732be7e9b10f3ff2e0ea60f67b89bc"} Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.515490 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="321f478ec960f5e41f07fb515a6eb67177732be7e9b10f3ff2e0ea60f67b89bc" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.515173 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-47zsh" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.572571 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.581428 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/de6c4b14-17ae-4701-b7bc-b4d907f6964c-etc-swift\") pod \"swift-storage-0\" (UID: \"de6c4b14-17ae-4701-b7bc-b4d907f6964c\") " pod="openstack/swift-storage-0" Dec 10 09:51:18 crc kubenswrapper[4715]: I1210 09:51:18.757629 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.041095 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-lz7jm" podUID="6fc17771-af47-4ba4-be8a-6b3143a999f4" containerName="ovn-controller" probeResult="failure" output=< Dec 10 09:51:19 crc kubenswrapper[4715]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 10 09:51:19 crc kubenswrapper[4715]: > Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.048161 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.057283 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rv2hq" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.215314 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 10 09:51:19 crc kubenswrapper[4715]: W1210 09:51:19.219875 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde6c4b14_17ae_4701_b7bc_b4d907f6964c.slice/crio-d830fb06459a59302c2e7ab8c5f1af57cd48c85715d5046dac51a2ccbccb4e5b WatchSource:0}: Error finding container d830fb06459a59302c2e7ab8c5f1af57cd48c85715d5046dac51a2ccbccb4e5b: Status 404 returned error can't find the container with id d830fb06459a59302c2e7ab8c5f1af57cd48c85715d5046dac51a2ccbccb4e5b Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.527155 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"d830fb06459a59302c2e7ab8c5f1af57cd48c85715d5046dac51a2ccbccb4e5b"} Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.530712 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b4bfd970-df85-4d59-a211-ba4adbb501ee","Type":"ContainerStarted","Data":"8c6195c9d724c71d578a4e356f8e2c517cd403d815c06e98ec486c70e110b6ce"} Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.550801 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"219b8c44-8f39-40f1-9000-25404c35d495","Type":"ContainerStarted","Data":"6760e3a41113c166d0e1e067ac2ea676271ab443db61e97d896a822f788a9fc9"} Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.555773 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=55.111065447 podStartE2EDuration="1m1.555755901s" podCreationTimestamp="2025-12-10 09:50:18 +0000 UTC" firstStartedPulling="2025-12-10 09:50:35.158832192 +0000 UTC m=+997.902378443" lastFinishedPulling="2025-12-10 09:50:41.603522646 +0000 UTC m=+1004.347068897" observedRunningTime="2025-12-10 09:51:19.551589523 +0000 UTC m=+1042.295135774" watchObservedRunningTime="2025-12-10 09:51:19.555755901 +0000 UTC m=+1042.299302152" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.638243 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=52.403633873 podStartE2EDuration="58.638220345s" podCreationTimestamp="2025-12-10 09:50:21 +0000 UTC" firstStartedPulling="2025-12-10 09:50:35.349630305 +0000 UTC m=+998.093176556" lastFinishedPulling="2025-12-10 09:50:41.584216767 +0000 UTC m=+1004.327763028" observedRunningTime="2025-12-10 09:51:19.60199575 +0000 UTC m=+1042.345542011" watchObservedRunningTime="2025-12-10 09:51:19.638220345 +0000 UTC m=+1042.381766596" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.888696 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 10 09:51:19 crc kubenswrapper[4715]: E1210 09:51:19.889116 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a351985-c466-4a2b-b74e-f67d85fb5715" containerName="swift-ring-rebalance" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.889137 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a351985-c466-4a2b-b74e-f67d85fb5715" containerName="swift-ring-rebalance" Dec 10 09:51:19 crc kubenswrapper[4715]: E1210 09:51:19.889170 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fbc1385-27ee-4c8b-8ce5-de03f14b4876" containerName="init" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.889179 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fbc1385-27ee-4c8b-8ce5-de03f14b4876" containerName="init" Dec 10 09:51:19 crc kubenswrapper[4715]: E1210 09:51:19.889202 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fbc1385-27ee-4c8b-8ce5-de03f14b4876" containerName="dnsmasq-dns" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.889211 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fbc1385-27ee-4c8b-8ce5-de03f14b4876" containerName="dnsmasq-dns" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.889397 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fbc1385-27ee-4c8b-8ce5-de03f14b4876" containerName="dnsmasq-dns" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.889414 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a351985-c466-4a2b-b74e-f67d85fb5715" containerName="swift-ring-rebalance" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.890475 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.900154 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.900388 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-9vhtl" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.900889 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.900978 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.918150 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.964564 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-lz7jm-config-z2pqf"] Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.967933 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.970066 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 10 09:51:19 crc kubenswrapper[4715]: I1210 09:51:19.983988 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-lz7jm-config-z2pqf"] Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.029000 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzgzn\" (UniqueName: \"kubernetes.io/projected/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-kube-api-access-hzgzn\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.029045 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-config\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.029083 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.029107 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.029127 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-scripts\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.029164 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.029194 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130144 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-log-ovn\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130195 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130238 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfz4v\" (UniqueName: \"kubernetes.io/projected/5086d94b-a9c9-41e2-9707-480b1775697a-kube-api-access-rfz4v\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130258 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-additional-scripts\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130289 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzgzn\" (UniqueName: \"kubernetes.io/projected/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-kube-api-access-hzgzn\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130313 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-config\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130345 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130360 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-scripts\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130382 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130401 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-scripts\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130437 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130468 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.130486 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run-ovn\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.131763 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-config\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.131775 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-scripts\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.132251 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.137031 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.139677 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.157583 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.191077 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzgzn\" (UniqueName: \"kubernetes.io/projected/1b0f6351-5ab5-4e0c-ad34-95061bc14a35-kube-api-access-hzgzn\") pod \"ovn-northd-0\" (UID: \"1b0f6351-5ab5-4e0c-ad34-95061bc14a35\") " pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.220529 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.232122 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfz4v\" (UniqueName: \"kubernetes.io/projected/5086d94b-a9c9-41e2-9707-480b1775697a-kube-api-access-rfz4v\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.232221 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-additional-scripts\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.232321 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-scripts\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.232471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run-ovn\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.232520 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-log-ovn\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.232544 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.233007 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.233104 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run-ovn\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.233181 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-log-ovn\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.234476 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-additional-scripts\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.235232 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-scripts\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.255056 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfz4v\" (UniqueName: \"kubernetes.io/projected/5086d94b-a9c9-41e2-9707-480b1775697a-kube-api-access-rfz4v\") pod \"ovn-controller-lz7jm-config-z2pqf\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.291546 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.374186 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.792684 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 10 09:51:20 crc kubenswrapper[4715]: W1210 09:51:20.796897 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b0f6351_5ab5_4e0c_ad34_95061bc14a35.slice/crio-46e349319767d6dda0cf19cae022401eb8d65a69a5c6bae212bb17d812c9869f WatchSource:0}: Error finding container 46e349319767d6dda0cf19cae022401eb8d65a69a5c6bae212bb17d812c9869f: Status 404 returned error can't find the container with id 46e349319767d6dda0cf19cae022401eb8d65a69a5c6bae212bb17d812c9869f Dec 10 09:51:20 crc kubenswrapper[4715]: I1210 09:51:20.881688 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-lz7jm-config-z2pqf"] Dec 10 09:51:21 crc kubenswrapper[4715]: W1210 09:51:21.104906 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5086d94b_a9c9_41e2_9707_480b1775697a.slice/crio-775f293785b4609923a00d563923304577870f481b6ff2e9f0d957f616acd8d3 WatchSource:0}: Error finding container 775f293785b4609923a00d563923304577870f481b6ff2e9f0d957f616acd8d3: Status 404 returned error can't find the container with id 775f293785b4609923a00d563923304577870f481b6ff2e9f0d957f616acd8d3 Dec 10 09:51:21 crc kubenswrapper[4715]: I1210 09:51:21.566139 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-lz7jm-config-z2pqf" event={"ID":"5086d94b-a9c9-41e2-9707-480b1775697a","Type":"ContainerStarted","Data":"4bc5cf13ef13df14c971a186ec8641765e3aa9c68ef0332d60027c065e9b6aa9"} Dec 10 09:51:21 crc kubenswrapper[4715]: I1210 09:51:21.566494 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-lz7jm-config-z2pqf" event={"ID":"5086d94b-a9c9-41e2-9707-480b1775697a","Type":"ContainerStarted","Data":"775f293785b4609923a00d563923304577870f481b6ff2e9f0d957f616acd8d3"} Dec 10 09:51:21 crc kubenswrapper[4715]: I1210 09:51:21.571622 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"44397680bed6a27f341a051de6a330632026da65a2d2e199c6367d0c6073a53c"} Dec 10 09:51:21 crc kubenswrapper[4715]: I1210 09:51:21.573064 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b0f6351-5ab5-4e0c-ad34-95061bc14a35","Type":"ContainerStarted","Data":"46e349319767d6dda0cf19cae022401eb8d65a69a5c6bae212bb17d812c9869f"} Dec 10 09:51:21 crc kubenswrapper[4715]: I1210 09:51:21.584152 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-lz7jm-config-z2pqf" podStartSLOduration=2.584136019 podStartE2EDuration="2.584136019s" podCreationTimestamp="2025-12-10 09:51:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:21.580769144 +0000 UTC m=+1044.324315395" watchObservedRunningTime="2025-12-10 09:51:21.584136019 +0000 UTC m=+1044.327682260" Dec 10 09:51:21 crc kubenswrapper[4715]: E1210 09:51:21.989176 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5086d94b_a9c9_41e2_9707_480b1775697a.slice/crio-4bc5cf13ef13df14c971a186ec8641765e3aa9c68ef0332d60027c065e9b6aa9.scope\": RecentStats: unable to find data in memory cache]" Dec 10 09:51:22 crc kubenswrapper[4715]: I1210 09:51:22.584315 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"043be8ce18e0d31137c0c09ee26450fb40f524782d8f42904a654b466d9ee6d7"} Dec 10 09:51:22 crc kubenswrapper[4715]: I1210 09:51:22.584661 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"fa86f920b4b256afd0b082e8729c722310050bbbebe8281190b5b67a28cc4f2b"} Dec 10 09:51:22 crc kubenswrapper[4715]: I1210 09:51:22.592184 4715 generic.go:334] "Generic (PLEG): container finished" podID="5086d94b-a9c9-41e2-9707-480b1775697a" containerID="4bc5cf13ef13df14c971a186ec8641765e3aa9c68ef0332d60027c065e9b6aa9" exitCode=0 Dec 10 09:51:22 crc kubenswrapper[4715]: I1210 09:51:22.592229 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-lz7jm-config-z2pqf" event={"ID":"5086d94b-a9c9-41e2-9707-480b1775697a","Type":"ContainerDied","Data":"4bc5cf13ef13df14c971a186ec8641765e3aa9c68ef0332d60027c065e9b6aa9"} Dec 10 09:51:24 crc kubenswrapper[4715]: I1210 09:51:24.049689 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-lz7jm" Dec 10 09:51:25 crc kubenswrapper[4715]: I1210 09:51:25.729444 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jzv2h" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerName="registry-server" probeResult="failure" output=< Dec 10 09:51:25 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 10 09:51:25 crc kubenswrapper[4715]: > Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.835663 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.969275 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run\") pod \"5086d94b-a9c9-41e2-9707-480b1775697a\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.969377 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run" (OuterVolumeSpecName: "var-run") pod "5086d94b-a9c9-41e2-9707-480b1775697a" (UID: "5086d94b-a9c9-41e2-9707-480b1775697a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.969415 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfz4v\" (UniqueName: \"kubernetes.io/projected/5086d94b-a9c9-41e2-9707-480b1775697a-kube-api-access-rfz4v\") pod \"5086d94b-a9c9-41e2-9707-480b1775697a\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.969516 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-log-ovn\") pod \"5086d94b-a9c9-41e2-9707-480b1775697a\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.969585 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run-ovn\") pod \"5086d94b-a9c9-41e2-9707-480b1775697a\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.969678 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-scripts\") pod \"5086d94b-a9c9-41e2-9707-480b1775697a\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.969733 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-additional-scripts\") pod \"5086d94b-a9c9-41e2-9707-480b1775697a\" (UID: \"5086d94b-a9c9-41e2-9707-480b1775697a\") " Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.969728 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "5086d94b-a9c9-41e2-9707-480b1775697a" (UID: "5086d94b-a9c9-41e2-9707-480b1775697a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.969726 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "5086d94b-a9c9-41e2-9707-480b1775697a" (UID: "5086d94b-a9c9-41e2-9707-480b1775697a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.970425 4715 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.970443 4715 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.970454 4715 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5086d94b-a9c9-41e2-9707-480b1775697a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.970888 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "5086d94b-a9c9-41e2-9707-480b1775697a" (UID: "5086d94b-a9c9-41e2-9707-480b1775697a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.971481 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-scripts" (OuterVolumeSpecName: "scripts") pod "5086d94b-a9c9-41e2-9707-480b1775697a" (UID: "5086d94b-a9c9-41e2-9707-480b1775697a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:29 crc kubenswrapper[4715]: I1210 09:51:29.975181 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5086d94b-a9c9-41e2-9707-480b1775697a-kube-api-access-rfz4v" (OuterVolumeSpecName: "kube-api-access-rfz4v") pod "5086d94b-a9c9-41e2-9707-480b1775697a" (UID: "5086d94b-a9c9-41e2-9707-480b1775697a"). InnerVolumeSpecName "kube-api-access-rfz4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.072547 4715 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.072572 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfz4v\" (UniqueName: \"kubernetes.io/projected/5086d94b-a9c9-41e2-9707-480b1775697a-kube-api-access-rfz4v\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.072585 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5086d94b-a9c9-41e2-9707-480b1775697a-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.379461 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.721525 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"fbd5cbc89af4ba069f0a201b8ed3ab51e70700f2b8dabf9551c8746ef90426a4"} Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.732381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b0f6351-5ab5-4e0c-ad34-95061bc14a35","Type":"ContainerStarted","Data":"922ff9ed37e6c5e652a761dce61dba9abe77448218892ed3eab22233faf2424d"} Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.732439 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b0f6351-5ab5-4e0c-ad34-95061bc14a35","Type":"ContainerStarted","Data":"19c14758332af145466695c0c569f2b239443dd42186ef872733af7573169a18"} Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.732473 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.736187 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-6h5ss"] Dec 10 09:51:30 crc kubenswrapper[4715]: E1210 09:51:30.736633 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5086d94b-a9c9-41e2-9707-480b1775697a" containerName="ovn-config" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.736656 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5086d94b-a9c9-41e2-9707-480b1775697a" containerName="ovn-config" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.736950 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5086d94b-a9c9-41e2-9707-480b1775697a" containerName="ovn-config" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.737671 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6h5ss" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.750586 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-lz7jm-config-z2pqf" event={"ID":"5086d94b-a9c9-41e2-9707-480b1775697a","Type":"ContainerDied","Data":"775f293785b4609923a00d563923304577870f481b6ff2e9f0d957f616acd8d3"} Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.750649 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="775f293785b4609923a00d563923304577870f481b6ff2e9f0d957f616acd8d3" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.750758 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-lz7jm-config-z2pqf" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.762934 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6h5ss"] Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.785136 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.802750 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.554563431 podStartE2EDuration="11.802728509s" podCreationTimestamp="2025-12-10 09:51:19 +0000 UTC" firstStartedPulling="2025-12-10 09:51:20.799475607 +0000 UTC m=+1043.543021848" lastFinishedPulling="2025-12-10 09:51:30.047640675 +0000 UTC m=+1052.791186926" observedRunningTime="2025-12-10 09:51:30.780768158 +0000 UTC m=+1053.524314409" watchObservedRunningTime="2025-12-10 09:51:30.802728509 +0000 UTC m=+1053.546274760" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.874672 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-x26bs"] Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.875736 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x26bs" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.889887 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-x26bs"] Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.901090 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fwcg\" (UniqueName: \"kubernetes.io/projected/cae7741b-9dd1-459a-a05e-c61a3c42b782-kube-api-access-8fwcg\") pod \"cinder-db-create-6h5ss\" (UID: \"cae7741b-9dd1-459a-a05e-c61a3c42b782\") " pod="openstack/cinder-db-create-6h5ss" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.901565 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae7741b-9dd1-459a-a05e-c61a3c42b782-operator-scripts\") pod \"cinder-db-create-6h5ss\" (UID: \"cae7741b-9dd1-459a-a05e-c61a3c42b782\") " pod="openstack/cinder-db-create-6h5ss" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.938971 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5098-account-create-update-n88lg"] Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.940199 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5098-account-create-update-n88lg" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.947207 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 10 09:51:30 crc kubenswrapper[4715]: I1210 09:51:30.958969 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5098-account-create-update-n88lg"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.002836 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae7741b-9dd1-459a-a05e-c61a3c42b782-operator-scripts\") pod \"cinder-db-create-6h5ss\" (UID: \"cae7741b-9dd1-459a-a05e-c61a3c42b782\") " pod="openstack/cinder-db-create-6h5ss" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.002928 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4ql5\" (UniqueName: \"kubernetes.io/projected/b0f80a84-53e6-4c55-89a0-fb526f3998c1-kube-api-access-j4ql5\") pod \"barbican-db-create-x26bs\" (UID: \"b0f80a84-53e6-4c55-89a0-fb526f3998c1\") " pod="openstack/barbican-db-create-x26bs" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.002982 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2228f25-b904-46ba-8999-23071a80094c-operator-scripts\") pod \"barbican-5098-account-create-update-n88lg\" (UID: \"b2228f25-b904-46ba-8999-23071a80094c\") " pod="openstack/barbican-5098-account-create-update-n88lg" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.003079 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f80a84-53e6-4c55-89a0-fb526f3998c1-operator-scripts\") pod \"barbican-db-create-x26bs\" (UID: \"b0f80a84-53e6-4c55-89a0-fb526f3998c1\") " pod="openstack/barbican-db-create-x26bs" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.003127 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fwcg\" (UniqueName: \"kubernetes.io/projected/cae7741b-9dd1-459a-a05e-c61a3c42b782-kube-api-access-8fwcg\") pod \"cinder-db-create-6h5ss\" (UID: \"cae7741b-9dd1-459a-a05e-c61a3c42b782\") " pod="openstack/cinder-db-create-6h5ss" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.003185 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7d5m\" (UniqueName: \"kubernetes.io/projected/b2228f25-b904-46ba-8999-23071a80094c-kube-api-access-p7d5m\") pod \"barbican-5098-account-create-update-n88lg\" (UID: \"b2228f25-b904-46ba-8999-23071a80094c\") " pod="openstack/barbican-5098-account-create-update-n88lg" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.003950 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae7741b-9dd1-459a-a05e-c61a3c42b782-operator-scripts\") pod \"cinder-db-create-6h5ss\" (UID: \"cae7741b-9dd1-459a-a05e-c61a3c42b782\") " pod="openstack/cinder-db-create-6h5ss" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.088958 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fwcg\" (UniqueName: \"kubernetes.io/projected/cae7741b-9dd1-459a-a05e-c61a3c42b782-kube-api-access-8fwcg\") pod \"cinder-db-create-6h5ss\" (UID: \"cae7741b-9dd1-459a-a05e-c61a3c42b782\") " pod="openstack/cinder-db-create-6h5ss" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.097124 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-lz7jm-config-z2pqf"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.104998 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4ql5\" (UniqueName: \"kubernetes.io/projected/b0f80a84-53e6-4c55-89a0-fb526f3998c1-kube-api-access-j4ql5\") pod \"barbican-db-create-x26bs\" (UID: \"b0f80a84-53e6-4c55-89a0-fb526f3998c1\") " pod="openstack/barbican-db-create-x26bs" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.105051 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2228f25-b904-46ba-8999-23071a80094c-operator-scripts\") pod \"barbican-5098-account-create-update-n88lg\" (UID: \"b2228f25-b904-46ba-8999-23071a80094c\") " pod="openstack/barbican-5098-account-create-update-n88lg" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.105116 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f80a84-53e6-4c55-89a0-fb526f3998c1-operator-scripts\") pod \"barbican-db-create-x26bs\" (UID: \"b0f80a84-53e6-4c55-89a0-fb526f3998c1\") " pod="openstack/barbican-db-create-x26bs" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.105159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7d5m\" (UniqueName: \"kubernetes.io/projected/b2228f25-b904-46ba-8999-23071a80094c-kube-api-access-p7d5m\") pod \"barbican-5098-account-create-update-n88lg\" (UID: \"b2228f25-b904-46ba-8999-23071a80094c\") " pod="openstack/barbican-5098-account-create-update-n88lg" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.106229 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2228f25-b904-46ba-8999-23071a80094c-operator-scripts\") pod \"barbican-5098-account-create-update-n88lg\" (UID: \"b2228f25-b904-46ba-8999-23071a80094c\") " pod="openstack/barbican-5098-account-create-update-n88lg" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.106346 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f80a84-53e6-4c55-89a0-fb526f3998c1-operator-scripts\") pod \"barbican-db-create-x26bs\" (UID: \"b0f80a84-53e6-4c55-89a0-fb526f3998c1\") " pod="openstack/barbican-db-create-x26bs" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.113353 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-lz7jm-config-z2pqf"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.131074 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6h5ss" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.166377 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7d5m\" (UniqueName: \"kubernetes.io/projected/b2228f25-b904-46ba-8999-23071a80094c-kube-api-access-p7d5m\") pod \"barbican-5098-account-create-update-n88lg\" (UID: \"b2228f25-b904-46ba-8999-23071a80094c\") " pod="openstack/barbican-5098-account-create-update-n88lg" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.175015 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4ql5\" (UniqueName: \"kubernetes.io/projected/b0f80a84-53e6-4c55-89a0-fb526f3998c1-kube-api-access-j4ql5\") pod \"barbican-db-create-x26bs\" (UID: \"b0f80a84-53e6-4c55-89a0-fb526f3998c1\") " pod="openstack/barbican-db-create-x26bs" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.227327 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x26bs" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.241616 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-38cd-account-create-update-pgrt5"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.245098 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-38cd-account-create-update-pgrt5" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.254876 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.267504 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5098-account-create-update-n88lg" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.268048 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-38cd-account-create-update-pgrt5"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.283237 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-bhnvd"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.284706 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.287895 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.289860 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-lz7jm-config-dnc8l"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.293943 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-csxkd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.294171 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.297247 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.303092 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.308074 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.308150 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6360fe51-a8bf-4b95-9d03-163543c7363e-operator-scripts\") pod \"cinder-38cd-account-create-update-pgrt5\" (UID: \"6360fe51-a8bf-4b95-9d03-163543c7363e\") " pod="openstack/cinder-38cd-account-create-update-pgrt5" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.308237 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hjjs\" (UniqueName: \"kubernetes.io/projected/6360fe51-a8bf-4b95-9d03-163543c7363e-kube-api-access-4hjjs\") pod \"cinder-38cd-account-create-update-pgrt5\" (UID: \"6360fe51-a8bf-4b95-9d03-163543c7363e\") " pod="openstack/cinder-38cd-account-create-update-pgrt5" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.310279 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bhnvd"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.337350 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-lz7jm-config-dnc8l"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.350992 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-68dkb"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.352040 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-68dkb" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.387520 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-68dkb"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410081 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-additional-scripts\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410125 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v9cs\" (UniqueName: \"kubernetes.io/projected/68b5f1b4-0e7f-4b10-808e-845621bc68d2-kube-api-access-9v9cs\") pod \"neutron-db-create-68dkb\" (UID: \"68b5f1b4-0e7f-4b10-808e-845621bc68d2\") " pod="openstack/neutron-db-create-68dkb" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410298 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b5f1b4-0e7f-4b10-808e-845621bc68d2-operator-scripts\") pod \"neutron-db-create-68dkb\" (UID: \"68b5f1b4-0e7f-4b10-808e-845621bc68d2\") " pod="openstack/neutron-db-create-68dkb" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410317 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-config-data\") pod \"keystone-db-sync-bhnvd\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410346 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hjjs\" (UniqueName: \"kubernetes.io/projected/6360fe51-a8bf-4b95-9d03-163543c7363e-kube-api-access-4hjjs\") pod \"cinder-38cd-account-create-update-pgrt5\" (UID: \"6360fe51-a8bf-4b95-9d03-163543c7363e\") " pod="openstack/cinder-38cd-account-create-update-pgrt5" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410386 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-log-ovn\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410409 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run-ovn\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410429 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-combined-ca-bundle\") pod \"keystone-db-sync-bhnvd\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410455 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-scripts\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410504 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twrwk\" (UniqueName: \"kubernetes.io/projected/1095391c-df23-4ac0-af2c-b3838d52874d-kube-api-access-twrwk\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410541 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6360fe51-a8bf-4b95-9d03-163543c7363e-operator-scripts\") pod \"cinder-38cd-account-create-update-pgrt5\" (UID: \"6360fe51-a8bf-4b95-9d03-163543c7363e\") " pod="openstack/cinder-38cd-account-create-update-pgrt5" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410660 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.410695 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkl6t\" (UniqueName: \"kubernetes.io/projected/108964f4-918d-4a16-93d2-85f0ee798efc-kube-api-access-gkl6t\") pod \"keystone-db-sync-bhnvd\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.411352 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6360fe51-a8bf-4b95-9d03-163543c7363e-operator-scripts\") pod \"cinder-38cd-account-create-update-pgrt5\" (UID: \"6360fe51-a8bf-4b95-9d03-163543c7363e\") " pod="openstack/cinder-38cd-account-create-update-pgrt5" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.438866 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hjjs\" (UniqueName: \"kubernetes.io/projected/6360fe51-a8bf-4b95-9d03-163543c7363e-kube-api-access-4hjjs\") pod \"cinder-38cd-account-create-update-pgrt5\" (UID: \"6360fe51-a8bf-4b95-9d03-163543c7363e\") " pod="openstack/cinder-38cd-account-create-update-pgrt5" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.511975 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twrwk\" (UniqueName: \"kubernetes.io/projected/1095391c-df23-4ac0-af2c-b3838d52874d-kube-api-access-twrwk\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.512294 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.512341 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkl6t\" (UniqueName: \"kubernetes.io/projected/108964f4-918d-4a16-93d2-85f0ee798efc-kube-api-access-gkl6t\") pod \"keystone-db-sync-bhnvd\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.512381 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-additional-scripts\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.512416 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v9cs\" (UniqueName: \"kubernetes.io/projected/68b5f1b4-0e7f-4b10-808e-845621bc68d2-kube-api-access-9v9cs\") pod \"neutron-db-create-68dkb\" (UID: \"68b5f1b4-0e7f-4b10-808e-845621bc68d2\") " pod="openstack/neutron-db-create-68dkb" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.512448 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b5f1b4-0e7f-4b10-808e-845621bc68d2-operator-scripts\") pod \"neutron-db-create-68dkb\" (UID: \"68b5f1b4-0e7f-4b10-808e-845621bc68d2\") " pod="openstack/neutron-db-create-68dkb" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.512471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-config-data\") pod \"keystone-db-sync-bhnvd\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.512515 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-log-ovn\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.512546 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run-ovn\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.512574 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-combined-ca-bundle\") pod \"keystone-db-sync-bhnvd\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.512607 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-scripts\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.513980 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.514487 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b5f1b4-0e7f-4b10-808e-845621bc68d2-operator-scripts\") pod \"neutron-db-create-68dkb\" (UID: \"68b5f1b4-0e7f-4b10-808e-845621bc68d2\") " pod="openstack/neutron-db-create-68dkb" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.514821 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-log-ovn\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.516486 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-additional-scripts\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.516557 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run-ovn\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.517624 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-scripts\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.532586 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-0bb5-account-create-update-7gr2k"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.534129 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0bb5-account-create-update-7gr2k" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.537534 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-0bb5-account-create-update-7gr2k"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.538389 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-config-data\") pod \"keystone-db-sync-bhnvd\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.539477 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.542463 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkl6t\" (UniqueName: \"kubernetes.io/projected/108964f4-918d-4a16-93d2-85f0ee798efc-kube-api-access-gkl6t\") pod \"keystone-db-sync-bhnvd\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.547704 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twrwk\" (UniqueName: \"kubernetes.io/projected/1095391c-df23-4ac0-af2c-b3838d52874d-kube-api-access-twrwk\") pod \"ovn-controller-lz7jm-config-dnc8l\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.550654 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-combined-ca-bundle\") pod \"keystone-db-sync-bhnvd\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.576056 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v9cs\" (UniqueName: \"kubernetes.io/projected/68b5f1b4-0e7f-4b10-808e-845621bc68d2-kube-api-access-9v9cs\") pod \"neutron-db-create-68dkb\" (UID: \"68b5f1b4-0e7f-4b10-808e-845621bc68d2\") " pod="openstack/neutron-db-create-68dkb" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.615501 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-operator-scripts\") pod \"neutron-0bb5-account-create-update-7gr2k\" (UID: \"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf\") " pod="openstack/neutron-0bb5-account-create-update-7gr2k" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.615678 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mm7j\" (UniqueName: \"kubernetes.io/projected/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-kube-api-access-6mm7j\") pod \"neutron-0bb5-account-create-update-7gr2k\" (UID: \"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf\") " pod="openstack/neutron-0bb5-account-create-update-7gr2k" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.627383 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5086d94b-a9c9-41e2-9707-480b1775697a" path="/var/lib/kubelet/pods/5086d94b-a9c9-41e2-9707-480b1775697a/volumes" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.633336 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-38cd-account-create-update-pgrt5" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.659728 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.677084 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.689602 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-68dkb" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.717935 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-operator-scripts\") pod \"neutron-0bb5-account-create-update-7gr2k\" (UID: \"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf\") " pod="openstack/neutron-0bb5-account-create-update-7gr2k" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.718198 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mm7j\" (UniqueName: \"kubernetes.io/projected/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-kube-api-access-6mm7j\") pod \"neutron-0bb5-account-create-update-7gr2k\" (UID: \"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf\") " pod="openstack/neutron-0bb5-account-create-update-7gr2k" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.719547 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-operator-scripts\") pod \"neutron-0bb5-account-create-update-7gr2k\" (UID: \"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf\") " pod="openstack/neutron-0bb5-account-create-update-7gr2k" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.737097 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mm7j\" (UniqueName: \"kubernetes.io/projected/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-kube-api-access-6mm7j\") pod \"neutron-0bb5-account-create-update-7gr2k\" (UID: \"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf\") " pod="openstack/neutron-0bb5-account-create-update-7gr2k" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.761078 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gd24z" event={"ID":"8ad144d0-9a73-4736-8f88-43dd295478d7","Type":"ContainerStarted","Data":"2e9be1f8f85f577f3901b5548949f35e2fe59a0f0b9e968f924ae21ca3f853fb"} Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.841689 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-gd24z" podStartSLOduration=3.720064722 podStartE2EDuration="18.841670609s" podCreationTimestamp="2025-12-10 09:51:13 +0000 UTC" firstStartedPulling="2025-12-10 09:51:14.960253406 +0000 UTC m=+1037.703799657" lastFinishedPulling="2025-12-10 09:51:30.081859303 +0000 UTC m=+1052.825405544" observedRunningTime="2025-12-10 09:51:31.783824401 +0000 UTC m=+1054.527370642" watchObservedRunningTime="2025-12-10 09:51:31.841670609 +0000 UTC m=+1054.585216860" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.847130 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6h5ss"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.870293 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0bb5-account-create-update-7gr2k" Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.963448 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5098-account-create-update-n88lg"] Dec 10 09:51:31 crc kubenswrapper[4715]: I1210 09:51:31.985906 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-x26bs"] Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.677427 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-68dkb"] Dec 10 09:51:32 crc kubenswrapper[4715]: W1210 09:51:32.682205 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68b5f1b4_0e7f_4b10_808e_845621bc68d2.slice/crio-030b716c087a2b6e1b1317d5b476eb9a37eecbaf68ba231c667e394df240df82 WatchSource:0}: Error finding container 030b716c087a2b6e1b1317d5b476eb9a37eecbaf68ba231c667e394df240df82: Status 404 returned error can't find the container with id 030b716c087a2b6e1b1317d5b476eb9a37eecbaf68ba231c667e394df240df82 Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.786332 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6h5ss" event={"ID":"cae7741b-9dd1-459a-a05e-c61a3c42b782","Type":"ContainerStarted","Data":"6ec18426bc83ba73a7fa8c4317f0d027df1e768fd368168af3230338213e4d0c"} Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.786558 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6h5ss" event={"ID":"cae7741b-9dd1-459a-a05e-c61a3c42b782","Type":"ContainerStarted","Data":"02c378d875d4fd169a2ec9b29dbd44604dc6ffe09d9093a17e35083a8566f576"} Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.787560 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-68dkb" event={"ID":"68b5f1b4-0e7f-4b10-808e-845621bc68d2","Type":"ContainerStarted","Data":"030b716c087a2b6e1b1317d5b476eb9a37eecbaf68ba231c667e394df240df82"} Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.788634 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x26bs" event={"ID":"b0f80a84-53e6-4c55-89a0-fb526f3998c1","Type":"ContainerStarted","Data":"f5f1356f1edc87ef4de08ea55fcefe3e1ec816f67580dd41112ccc4c4ccab24e"} Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.788736 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x26bs" event={"ID":"b0f80a84-53e6-4c55-89a0-fb526f3998c1","Type":"ContainerStarted","Data":"a9729d0d059a8a8dd12b08152ab8047a2f50aeb6e1e469df098c3c3f0027c459"} Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.794783 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5098-account-create-update-n88lg" event={"ID":"b2228f25-b904-46ba-8999-23071a80094c","Type":"ContainerStarted","Data":"9e96cdcfdabf254db79354a432fcf5f570a5edf6123cbce6e6057b52ace584df"} Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.795035 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5098-account-create-update-n88lg" event={"ID":"b2228f25-b904-46ba-8999-23071a80094c","Type":"ContainerStarted","Data":"5a1251182da3d8b537f577e16e66ffdb521b7331f0c5b793409cec66eb89ee3c"} Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.819005 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-38cd-account-create-update-pgrt5"] Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.851565 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bhnvd"] Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.856664 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-x26bs" podStartSLOduration=2.8566446599999997 podStartE2EDuration="2.85664466s" podCreationTimestamp="2025-12-10 09:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:32.825759986 +0000 UTC m=+1055.569306237" watchObservedRunningTime="2025-12-10 09:51:32.85664466 +0000 UTC m=+1055.600190911" Dec 10 09:51:32 crc kubenswrapper[4715]: I1210 09:51:32.866402 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-5098-account-create-update-n88lg" podStartSLOduration=2.8663796059999997 podStartE2EDuration="2.866379606s" podCreationTimestamp="2025-12-10 09:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:32.845364431 +0000 UTC m=+1055.588910692" watchObservedRunningTime="2025-12-10 09:51:32.866379606 +0000 UTC m=+1055.609925857" Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.019690 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-0bb5-account-create-update-7gr2k"] Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.056418 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-lz7jm-config-dnc8l"] Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.823286 4715 generic.go:334] "Generic (PLEG): container finished" podID="cae7741b-9dd1-459a-a05e-c61a3c42b782" containerID="6ec18426bc83ba73a7fa8c4317f0d027df1e768fd368168af3230338213e4d0c" exitCode=0 Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.823405 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6h5ss" event={"ID":"cae7741b-9dd1-459a-a05e-c61a3c42b782","Type":"ContainerDied","Data":"6ec18426bc83ba73a7fa8c4317f0d027df1e768fd368168af3230338213e4d0c"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.829864 4715 generic.go:334] "Generic (PLEG): container finished" podID="68b5f1b4-0e7f-4b10-808e-845621bc68d2" containerID="2405ad40366ebf0445c1b61deba023d87a8460937c11c891e3a3cb9a30b1b503" exitCode=0 Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.829966 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-68dkb" event={"ID":"68b5f1b4-0e7f-4b10-808e-845621bc68d2","Type":"ContainerDied","Data":"2405ad40366ebf0445c1b61deba023d87a8460937c11c891e3a3cb9a30b1b503"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.832399 4715 generic.go:334] "Generic (PLEG): container finished" podID="b2228f25-b904-46ba-8999-23071a80094c" containerID="9e96cdcfdabf254db79354a432fcf5f570a5edf6123cbce6e6057b52ace584df" exitCode=0 Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.832438 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5098-account-create-update-n88lg" event={"ID":"b2228f25-b904-46ba-8999-23071a80094c","Type":"ContainerDied","Data":"9e96cdcfdabf254db79354a432fcf5f570a5edf6123cbce6e6057b52ace584df"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.835693 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-lz7jm-config-dnc8l" event={"ID":"1095391c-df23-4ac0-af2c-b3838d52874d","Type":"ContainerStarted","Data":"7fe183b9869687758f0327fc261359e2806380711d2ad774c34e9dcf7fc8424a"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.835752 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-lz7jm-config-dnc8l" event={"ID":"1095391c-df23-4ac0-af2c-b3838d52874d","Type":"ContainerStarted","Data":"94cfd85d90ec702e605ffc2290e34e047eb07e8950831c9550ff3ef71e804b2b"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.842678 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"83a087c501fa304ff64297c375822ed73d564b24f10913c4003541303bcb6317"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.842720 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"c25d5551398e0a3c98a3a75d44e642a8c66b91ff93dfc6a8464b006983a62008"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.844581 4715 generic.go:334] "Generic (PLEG): container finished" podID="b0f80a84-53e6-4c55-89a0-fb526f3998c1" containerID="f5f1356f1edc87ef4de08ea55fcefe3e1ec816f67580dd41112ccc4c4ccab24e" exitCode=0 Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.844644 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x26bs" event={"ID":"b0f80a84-53e6-4c55-89a0-fb526f3998c1","Type":"ContainerDied","Data":"f5f1356f1edc87ef4de08ea55fcefe3e1ec816f67580dd41112ccc4c4ccab24e"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.878736 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0bb5-account-create-update-7gr2k" event={"ID":"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf","Type":"ContainerStarted","Data":"c994c59c3996db0033fcbfd436c5edf439427180f8eb91692bfb34864bf118b9"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.878785 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0bb5-account-create-update-7gr2k" event={"ID":"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf","Type":"ContainerStarted","Data":"35e9dfb9642b29c9dff573be1116a81655206ce0e7c3aa664b82f8436eb3843b"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.880566 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bhnvd" event={"ID":"108964f4-918d-4a16-93d2-85f0ee798efc","Type":"ContainerStarted","Data":"ad30b411c552345ad198331c1400ecd282fe9de56f3f58afe53e9daecc302f8f"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.885378 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-38cd-account-create-update-pgrt5" event={"ID":"6360fe51-a8bf-4b95-9d03-163543c7363e","Type":"ContainerStarted","Data":"6292a5055b87b5d45f479cf639b29d193fb29a445d5e3d584134842a70af9e8a"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.885443 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-38cd-account-create-update-pgrt5" event={"ID":"6360fe51-a8bf-4b95-9d03-163543c7363e","Type":"ContainerStarted","Data":"ddfb0ddcad5828b69ec6d18c9905502c1b6748ce1c8617b1bf0ae3430c9dc036"} Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.920089 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-lz7jm-config-dnc8l" podStartSLOduration=2.9200718119999998 podStartE2EDuration="2.920071812s" podCreationTimestamp="2025-12-10 09:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:33.909430961 +0000 UTC m=+1056.652977222" watchObservedRunningTime="2025-12-10 09:51:33.920071812 +0000 UTC m=+1056.663618063" Dec 10 09:51:33 crc kubenswrapper[4715]: I1210 09:51:33.973204 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-38cd-account-create-update-pgrt5" podStartSLOduration=2.973188415 podStartE2EDuration="2.973188415s" podCreationTimestamp="2025-12-10 09:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:33.967574606 +0000 UTC m=+1056.711120857" watchObservedRunningTime="2025-12-10 09:51:33.973188415 +0000 UTC m=+1056.716734666" Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.006853 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-0bb5-account-create-update-7gr2k" podStartSLOduration=3.006837708 podStartE2EDuration="3.006837708s" podCreationTimestamp="2025-12-10 09:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:33.997698769 +0000 UTC m=+1056.741245020" watchObservedRunningTime="2025-12-10 09:51:34.006837708 +0000 UTC m=+1056.750383959" Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.732678 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.780461 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.897906 4715 generic.go:334] "Generic (PLEG): container finished" podID="6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf" containerID="c994c59c3996db0033fcbfd436c5edf439427180f8eb91692bfb34864bf118b9" exitCode=0 Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.898007 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0bb5-account-create-update-7gr2k" event={"ID":"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf","Type":"ContainerDied","Data":"c994c59c3996db0033fcbfd436c5edf439427180f8eb91692bfb34864bf118b9"} Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.900877 4715 generic.go:334] "Generic (PLEG): container finished" podID="6360fe51-a8bf-4b95-9d03-163543c7363e" containerID="6292a5055b87b5d45f479cf639b29d193fb29a445d5e3d584134842a70af9e8a" exitCode=0 Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.900983 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-38cd-account-create-update-pgrt5" event={"ID":"6360fe51-a8bf-4b95-9d03-163543c7363e","Type":"ContainerDied","Data":"6292a5055b87b5d45f479cf639b29d193fb29a445d5e3d584134842a70af9e8a"} Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.903165 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-lz7jm-config-dnc8l" event={"ID":"1095391c-df23-4ac0-af2c-b3838d52874d","Type":"ContainerDied","Data":"7fe183b9869687758f0327fc261359e2806380711d2ad774c34e9dcf7fc8424a"} Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.905451 4715 generic.go:334] "Generic (PLEG): container finished" podID="1095391c-df23-4ac0-af2c-b3838d52874d" containerID="7fe183b9869687758f0327fc261359e2806380711d2ad774c34e9dcf7fc8424a" exitCode=0 Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.913781 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"517dd28bbd2b007a3c3e1520ac604a3c12b74ac0ff53327af961d08b8846ca9a"} Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.913822 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"62b69a307a6a2aca02bdce09c68dfdf9f2adefe486a33773485cea39643f87fe"} Dec 10 09:51:34 crc kubenswrapper[4715]: I1210 09:51:34.983772 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzv2h"] Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.371145 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-68dkb" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.402798 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v9cs\" (UniqueName: \"kubernetes.io/projected/68b5f1b4-0e7f-4b10-808e-845621bc68d2-kube-api-access-9v9cs\") pod \"68b5f1b4-0e7f-4b10-808e-845621bc68d2\" (UID: \"68b5f1b4-0e7f-4b10-808e-845621bc68d2\") " Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.403005 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b5f1b4-0e7f-4b10-808e-845621bc68d2-operator-scripts\") pod \"68b5f1b4-0e7f-4b10-808e-845621bc68d2\" (UID: \"68b5f1b4-0e7f-4b10-808e-845621bc68d2\") " Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.404097 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68b5f1b4-0e7f-4b10-808e-845621bc68d2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "68b5f1b4-0e7f-4b10-808e-845621bc68d2" (UID: "68b5f1b4-0e7f-4b10-808e-845621bc68d2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.410863 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b5f1b4-0e7f-4b10-808e-845621bc68d2-kube-api-access-9v9cs" (OuterVolumeSpecName: "kube-api-access-9v9cs") pod "68b5f1b4-0e7f-4b10-808e-845621bc68d2" (UID: "68b5f1b4-0e7f-4b10-808e-845621bc68d2"). InnerVolumeSpecName "kube-api-access-9v9cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.496842 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5098-account-create-update-n88lg" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.504730 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68b5f1b4-0e7f-4b10-808e-845621bc68d2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.504766 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v9cs\" (UniqueName: \"kubernetes.io/projected/68b5f1b4-0e7f-4b10-808e-845621bc68d2-kube-api-access-9v9cs\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.506623 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x26bs" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.516343 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6h5ss" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.605636 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7d5m\" (UniqueName: \"kubernetes.io/projected/b2228f25-b904-46ba-8999-23071a80094c-kube-api-access-p7d5m\") pod \"b2228f25-b904-46ba-8999-23071a80094c\" (UID: \"b2228f25-b904-46ba-8999-23071a80094c\") " Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.605752 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f80a84-53e6-4c55-89a0-fb526f3998c1-operator-scripts\") pod \"b0f80a84-53e6-4c55-89a0-fb526f3998c1\" (UID: \"b0f80a84-53e6-4c55-89a0-fb526f3998c1\") " Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.605779 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2228f25-b904-46ba-8999-23071a80094c-operator-scripts\") pod \"b2228f25-b904-46ba-8999-23071a80094c\" (UID: \"b2228f25-b904-46ba-8999-23071a80094c\") " Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.606057 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4ql5\" (UniqueName: \"kubernetes.io/projected/b0f80a84-53e6-4c55-89a0-fb526f3998c1-kube-api-access-j4ql5\") pod \"b0f80a84-53e6-4c55-89a0-fb526f3998c1\" (UID: \"b0f80a84-53e6-4c55-89a0-fb526f3998c1\") " Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.606082 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae7741b-9dd1-459a-a05e-c61a3c42b782-operator-scripts\") pod \"cae7741b-9dd1-459a-a05e-c61a3c42b782\" (UID: \"cae7741b-9dd1-459a-a05e-c61a3c42b782\") " Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.606490 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f80a84-53e6-4c55-89a0-fb526f3998c1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0f80a84-53e6-4c55-89a0-fb526f3998c1" (UID: "b0f80a84-53e6-4c55-89a0-fb526f3998c1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.606689 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2228f25-b904-46ba-8999-23071a80094c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b2228f25-b904-46ba-8999-23071a80094c" (UID: "b2228f25-b904-46ba-8999-23071a80094c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.606809 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fwcg\" (UniqueName: \"kubernetes.io/projected/cae7741b-9dd1-459a-a05e-c61a3c42b782-kube-api-access-8fwcg\") pod \"cae7741b-9dd1-459a-a05e-c61a3c42b782\" (UID: \"cae7741b-9dd1-459a-a05e-c61a3c42b782\") " Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.607003 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cae7741b-9dd1-459a-a05e-c61a3c42b782-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cae7741b-9dd1-459a-a05e-c61a3c42b782" (UID: "cae7741b-9dd1-459a-a05e-c61a3c42b782"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.607768 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f80a84-53e6-4c55-89a0-fb526f3998c1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.607794 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2228f25-b904-46ba-8999-23071a80094c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.607807 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae7741b-9dd1-459a-a05e-c61a3c42b782-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.610226 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0f80a84-53e6-4c55-89a0-fb526f3998c1-kube-api-access-j4ql5" (OuterVolumeSpecName: "kube-api-access-j4ql5") pod "b0f80a84-53e6-4c55-89a0-fb526f3998c1" (UID: "b0f80a84-53e6-4c55-89a0-fb526f3998c1"). InnerVolumeSpecName "kube-api-access-j4ql5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.610631 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cae7741b-9dd1-459a-a05e-c61a3c42b782-kube-api-access-8fwcg" (OuterVolumeSpecName: "kube-api-access-8fwcg") pod "cae7741b-9dd1-459a-a05e-c61a3c42b782" (UID: "cae7741b-9dd1-459a-a05e-c61a3c42b782"). InnerVolumeSpecName "kube-api-access-8fwcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.614741 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2228f25-b904-46ba-8999-23071a80094c-kube-api-access-p7d5m" (OuterVolumeSpecName: "kube-api-access-p7d5m") pod "b2228f25-b904-46ba-8999-23071a80094c" (UID: "b2228f25-b904-46ba-8999-23071a80094c"). InnerVolumeSpecName "kube-api-access-p7d5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.709720 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7d5m\" (UniqueName: \"kubernetes.io/projected/b2228f25-b904-46ba-8999-23071a80094c-kube-api-access-p7d5m\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.709765 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4ql5\" (UniqueName: \"kubernetes.io/projected/b0f80a84-53e6-4c55-89a0-fb526f3998c1-kube-api-access-j4ql5\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.709778 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fwcg\" (UniqueName: \"kubernetes.io/projected/cae7741b-9dd1-459a-a05e-c61a3c42b782-kube-api-access-8fwcg\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.927139 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6h5ss" event={"ID":"cae7741b-9dd1-459a-a05e-c61a3c42b782","Type":"ContainerDied","Data":"02c378d875d4fd169a2ec9b29dbd44604dc6ffe09d9093a17e35083a8566f576"} Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.927177 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02c378d875d4fd169a2ec9b29dbd44604dc6ffe09d9093a17e35083a8566f576" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.927228 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6h5ss" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.931709 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-68dkb" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.932115 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-68dkb" event={"ID":"68b5f1b4-0e7f-4b10-808e-845621bc68d2","Type":"ContainerDied","Data":"030b716c087a2b6e1b1317d5b476eb9a37eecbaf68ba231c667e394df240df82"} Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.932137 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="030b716c087a2b6e1b1317d5b476eb9a37eecbaf68ba231c667e394df240df82" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.934372 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x26bs" event={"ID":"b0f80a84-53e6-4c55-89a0-fb526f3998c1","Type":"ContainerDied","Data":"a9729d0d059a8a8dd12b08152ab8047a2f50aeb6e1e469df098c3c3f0027c459"} Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.934403 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9729d0d059a8a8dd12b08152ab8047a2f50aeb6e1e469df098c3c3f0027c459" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.934449 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x26bs" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.938412 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5098-account-create-update-n88lg" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.939152 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5098-account-create-update-n88lg" event={"ID":"b2228f25-b904-46ba-8999-23071a80094c","Type":"ContainerDied","Data":"5a1251182da3d8b537f577e16e66ffdb521b7331f0c5b793409cec66eb89ee3c"} Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.939220 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a1251182da3d8b537f577e16e66ffdb521b7331f0c5b793409cec66eb89ee3c" Dec 10 09:51:35 crc kubenswrapper[4715]: I1210 09:51:35.940134 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jzv2h" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerName="registry-server" containerID="cri-o://d4b28f8243a1f77d4aecaad28451bb4efb5696b8f113462baccda860b55b1340" gracePeriod=2 Dec 10 09:51:36 crc kubenswrapper[4715]: I1210 09:51:36.952704 4715 generic.go:334] "Generic (PLEG): container finished" podID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerID="d4b28f8243a1f77d4aecaad28451bb4efb5696b8f113462baccda860b55b1340" exitCode=0 Dec 10 09:51:36 crc kubenswrapper[4715]: I1210 09:51:36.953169 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzv2h" event={"ID":"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c","Type":"ContainerDied","Data":"d4b28f8243a1f77d4aecaad28451bb4efb5696b8f113462baccda860b55b1340"} Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.569848 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-38cd-account-create-update-pgrt5" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.622179 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.636347 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0bb5-account-create-update-7gr2k" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666424 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-scripts\") pod \"1095391c-df23-4ac0-af2c-b3838d52874d\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666466 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run\") pod \"1095391c-df23-4ac0-af2c-b3838d52874d\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666526 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mm7j\" (UniqueName: \"kubernetes.io/projected/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-kube-api-access-6mm7j\") pod \"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf\" (UID: \"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666570 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-operator-scripts\") pod \"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf\" (UID: \"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666568 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run" (OuterVolumeSpecName: "var-run") pod "1095391c-df23-4ac0-af2c-b3838d52874d" (UID: "1095391c-df23-4ac0-af2c-b3838d52874d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666598 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run-ovn\") pod \"1095391c-df23-4ac0-af2c-b3838d52874d\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666633 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-log-ovn\") pod \"1095391c-df23-4ac0-af2c-b3838d52874d\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666686 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hjjs\" (UniqueName: \"kubernetes.io/projected/6360fe51-a8bf-4b95-9d03-163543c7363e-kube-api-access-4hjjs\") pod \"6360fe51-a8bf-4b95-9d03-163543c7363e\" (UID: \"6360fe51-a8bf-4b95-9d03-163543c7363e\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666763 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twrwk\" (UniqueName: \"kubernetes.io/projected/1095391c-df23-4ac0-af2c-b3838d52874d-kube-api-access-twrwk\") pod \"1095391c-df23-4ac0-af2c-b3838d52874d\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666795 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-additional-scripts\") pod \"1095391c-df23-4ac0-af2c-b3838d52874d\" (UID: \"1095391c-df23-4ac0-af2c-b3838d52874d\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666842 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6360fe51-a8bf-4b95-9d03-163543c7363e-operator-scripts\") pod \"6360fe51-a8bf-4b95-9d03-163543c7363e\" (UID: \"6360fe51-a8bf-4b95-9d03-163543c7363e\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.666979 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1095391c-df23-4ac0-af2c-b3838d52874d" (UID: "1095391c-df23-4ac0-af2c-b3838d52874d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.667272 4715 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.667290 4715 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.667676 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6360fe51-a8bf-4b95-9d03-163543c7363e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6360fe51-a8bf-4b95-9d03-163543c7363e" (UID: "6360fe51-a8bf-4b95-9d03-163543c7363e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.668013 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf" (UID: "6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.668274 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1095391c-df23-4ac0-af2c-b3838d52874d" (UID: "1095391c-df23-4ac0-af2c-b3838d52874d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.668362 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1095391c-df23-4ac0-af2c-b3838d52874d" (UID: "1095391c-df23-4ac0-af2c-b3838d52874d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.669031 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-scripts" (OuterVolumeSpecName: "scripts") pod "1095391c-df23-4ac0-af2c-b3838d52874d" (UID: "1095391c-df23-4ac0-af2c-b3838d52874d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.678087 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6360fe51-a8bf-4b95-9d03-163543c7363e-kube-api-access-4hjjs" (OuterVolumeSpecName: "kube-api-access-4hjjs") pod "6360fe51-a8bf-4b95-9d03-163543c7363e" (UID: "6360fe51-a8bf-4b95-9d03-163543c7363e"). InnerVolumeSpecName "kube-api-access-4hjjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.680077 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1095391c-df23-4ac0-af2c-b3838d52874d-kube-api-access-twrwk" (OuterVolumeSpecName: "kube-api-access-twrwk") pod "1095391c-df23-4ac0-af2c-b3838d52874d" (UID: "1095391c-df23-4ac0-af2c-b3838d52874d"). InnerVolumeSpecName "kube-api-access-twrwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.680116 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-kube-api-access-6mm7j" (OuterVolumeSpecName: "kube-api-access-6mm7j") pod "6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf" (UID: "6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf"). InnerVolumeSpecName "kube-api-access-6mm7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.768318 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mm7j\" (UniqueName: \"kubernetes.io/projected/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-kube-api-access-6mm7j\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.768357 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.768372 4715 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1095391c-df23-4ac0-af2c-b3838d52874d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.768384 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hjjs\" (UniqueName: \"kubernetes.io/projected/6360fe51-a8bf-4b95-9d03-163543c7363e-kube-api-access-4hjjs\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.768395 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twrwk\" (UniqueName: \"kubernetes.io/projected/1095391c-df23-4ac0-af2c-b3838d52874d-kube-api-access-twrwk\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.768405 4715 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.768418 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6360fe51-a8bf-4b95-9d03-163543c7363e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.768428 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1095391c-df23-4ac0-af2c-b3838d52874d-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.888076 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.972049 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-catalog-content\") pod \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.972360 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2jzf\" (UniqueName: \"kubernetes.io/projected/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-kube-api-access-r2jzf\") pod \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.972415 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-utilities\") pod \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\" (UID: \"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c\") " Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.973746 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-38cd-account-create-update-pgrt5" event={"ID":"6360fe51-a8bf-4b95-9d03-163543c7363e","Type":"ContainerDied","Data":"ddfb0ddcad5828b69ec6d18c9905502c1b6748ce1c8617b1bf0ae3430c9dc036"} Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.973782 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddfb0ddcad5828b69ec6d18c9905502c1b6748ce1c8617b1bf0ae3430c9dc036" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.973787 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-utilities" (OuterVolumeSpecName: "utilities") pod "01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" (UID: "01e46c2b-bbc9-46e8-bf17-ead30adb0f2c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.973851 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-38cd-account-create-update-pgrt5" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.976500 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-kube-api-access-r2jzf" (OuterVolumeSpecName: "kube-api-access-r2jzf") pod "01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" (UID: "01e46c2b-bbc9-46e8-bf17-ead30adb0f2c"). InnerVolumeSpecName "kube-api-access-r2jzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.993880 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzv2h" event={"ID":"01e46c2b-bbc9-46e8-bf17-ead30adb0f2c","Type":"ContainerDied","Data":"ac92fff7bc1fdfb72b1e54b64f35022af9d4e197bd50b9586fb97002aa3b1f2a"} Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.993960 4715 scope.go:117] "RemoveContainer" containerID="d4b28f8243a1f77d4aecaad28451bb4efb5696b8f113462baccda860b55b1340" Dec 10 09:51:38 crc kubenswrapper[4715]: I1210 09:51:38.994044 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzv2h" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.000585 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-lz7jm-config-dnc8l" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.000862 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-lz7jm-config-dnc8l" event={"ID":"1095391c-df23-4ac0-af2c-b3838d52874d","Type":"ContainerDied","Data":"94cfd85d90ec702e605ffc2290e34e047eb07e8950831c9550ff3ef71e804b2b"} Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.000893 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94cfd85d90ec702e605ffc2290e34e047eb07e8950831c9550ff3ef71e804b2b" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.024285 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"091e2fae2c240649d8c3318b0eaa8d44a43226b58ebd773466d1a0c4c441416d"} Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.026645 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0bb5-account-create-update-7gr2k" event={"ID":"6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf","Type":"ContainerDied","Data":"35e9dfb9642b29c9dff573be1116a81655206ce0e7c3aa664b82f8436eb3843b"} Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.026692 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35e9dfb9642b29c9dff573be1116a81655206ce0e7c3aa664b82f8436eb3843b" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.026748 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0bb5-account-create-update-7gr2k" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.043352 4715 scope.go:117] "RemoveContainer" containerID="0d66b62445cac79b9e76080cdfc5b4d9e3951ede5a1289448badd387bce34d52" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.071393 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" (UID: "01e46c2b-bbc9-46e8-bf17-ead30adb0f2c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.074877 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.074915 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2jzf\" (UniqueName: \"kubernetes.io/projected/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-kube-api-access-r2jzf\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.074937 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.093663 4715 scope.go:117] "RemoveContainer" containerID="9a8b667d87a402e885f86907168ff41d5b7ab2b45e1c533664c49c5ae49ae013" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.501017 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzv2h"] Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.506430 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jzv2h"] Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.619074 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" path="/var/lib/kubelet/pods/01e46c2b-bbc9-46e8-bf17-ead30adb0f2c/volumes" Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.721352 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-lz7jm-config-dnc8l"] Dec 10 09:51:39 crc kubenswrapper[4715]: I1210 09:51:39.730859 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-lz7jm-config-dnc8l"] Dec 10 09:51:40 crc kubenswrapper[4715]: I1210 09:51:40.040188 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"73df1c45b95da2aee4bbc9b729df5d6551cb93f6bb0c7ab0302b3d505ba911c4"} Dec 10 09:51:40 crc kubenswrapper[4715]: I1210 09:51:40.040238 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"e02b164f17e63a11fc2d1359d0920f62ea3d76a9d52a7087cd8e8c785968972e"} Dec 10 09:51:40 crc kubenswrapper[4715]: I1210 09:51:40.040251 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"3c5435dbb3ce0004fd0ee6242502f2ada7a634cf71f2323eb80c6b99f4eca8ba"} Dec 10 09:51:40 crc kubenswrapper[4715]: I1210 09:51:40.040261 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"bd74f7f214eac9d76dc3c539848259d9137bda1923647ba682c028967b4f2c05"} Dec 10 09:51:40 crc kubenswrapper[4715]: I1210 09:51:40.042315 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bhnvd" event={"ID":"108964f4-918d-4a16-93d2-85f0ee798efc","Type":"ContainerStarted","Data":"e68e45416c5c46e48d36f198c2ea5aa1081bf6d0dbc2a1e9a18ad82f2c64f29f"} Dec 10 09:51:40 crc kubenswrapper[4715]: I1210 09:51:40.068902 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-bhnvd" podStartSLOduration=3.2851993090000002 podStartE2EDuration="9.068882807s" podCreationTimestamp="2025-12-10 09:51:31 +0000 UTC" firstStartedPulling="2025-12-10 09:51:33.311715702 +0000 UTC m=+1056.055261953" lastFinishedPulling="2025-12-10 09:51:39.09539921 +0000 UTC m=+1061.838945451" observedRunningTime="2025-12-10 09:51:40.065947174 +0000 UTC m=+1062.809493435" watchObservedRunningTime="2025-12-10 09:51:40.068882807 +0000 UTC m=+1062.812429058" Dec 10 09:51:40 crc kubenswrapper[4715]: I1210 09:51:40.285354 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.053910 4715 generic.go:334] "Generic (PLEG): container finished" podID="8ad144d0-9a73-4736-8f88-43dd295478d7" containerID="2e9be1f8f85f577f3901b5548949f35e2fe59a0f0b9e968f924ae21ca3f853fb" exitCode=0 Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.054062 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gd24z" event={"ID":"8ad144d0-9a73-4736-8f88-43dd295478d7","Type":"ContainerDied","Data":"2e9be1f8f85f577f3901b5548949f35e2fe59a0f0b9e968f924ae21ca3f853fb"} Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.060237 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"321a7ca62c343b1a7a4f70cb5d7d7e06074f5eb2e173fe7309567222f6edbcb2"} Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.060291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"de6c4b14-17ae-4701-b7bc-b4d907f6964c","Type":"ContainerStarted","Data":"0ba1f8215ba01e50466e2636da9c73cdd3f70843a665a4ca01a5d31978d6e89a"} Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.113081 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.844834998 podStartE2EDuration="56.113060924s" podCreationTimestamp="2025-12-10 09:50:45 +0000 UTC" firstStartedPulling="2025-12-10 09:51:19.222082866 +0000 UTC m=+1041.965629117" lastFinishedPulling="2025-12-10 09:51:38.490308782 +0000 UTC m=+1061.233855043" observedRunningTime="2025-12-10 09:51:41.111219752 +0000 UTC m=+1063.854766043" watchObservedRunningTime="2025-12-10 09:51:41.113060924 +0000 UTC m=+1063.856607175" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.471122 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-tvr8x"] Dec 10 09:51:41 crc kubenswrapper[4715]: E1210 09:51:41.471784 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2228f25-b904-46ba-8999-23071a80094c" containerName="mariadb-account-create-update" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.471865 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2228f25-b904-46ba-8999-23071a80094c" containerName="mariadb-account-create-update" Dec 10 09:51:41 crc kubenswrapper[4715]: E1210 09:51:41.471952 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6360fe51-a8bf-4b95-9d03-163543c7363e" containerName="mariadb-account-create-update" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.472020 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6360fe51-a8bf-4b95-9d03-163543c7363e" containerName="mariadb-account-create-update" Dec 10 09:51:41 crc kubenswrapper[4715]: E1210 09:51:41.472096 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae7741b-9dd1-459a-a05e-c61a3c42b782" containerName="mariadb-database-create" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.472150 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae7741b-9dd1-459a-a05e-c61a3c42b782" containerName="mariadb-database-create" Dec 10 09:51:41 crc kubenswrapper[4715]: E1210 09:51:41.472220 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1095391c-df23-4ac0-af2c-b3838d52874d" containerName="ovn-config" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.472281 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1095391c-df23-4ac0-af2c-b3838d52874d" containerName="ovn-config" Dec 10 09:51:41 crc kubenswrapper[4715]: E1210 09:51:41.472340 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f80a84-53e6-4c55-89a0-fb526f3998c1" containerName="mariadb-database-create" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.472402 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f80a84-53e6-4c55-89a0-fb526f3998c1" containerName="mariadb-database-create" Dec 10 09:51:41 crc kubenswrapper[4715]: E1210 09:51:41.472461 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf" containerName="mariadb-account-create-update" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.472529 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf" containerName="mariadb-account-create-update" Dec 10 09:51:41 crc kubenswrapper[4715]: E1210 09:51:41.472587 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b5f1b4-0e7f-4b10-808e-845621bc68d2" containerName="mariadb-database-create" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.472649 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b5f1b4-0e7f-4b10-808e-845621bc68d2" containerName="mariadb-database-create" Dec 10 09:51:41 crc kubenswrapper[4715]: E1210 09:51:41.472730 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerName="extract-utilities" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.472797 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerName="extract-utilities" Dec 10 09:51:41 crc kubenswrapper[4715]: E1210 09:51:41.472900 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerName="extract-content" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.473002 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerName="extract-content" Dec 10 09:51:41 crc kubenswrapper[4715]: E1210 09:51:41.473121 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerName="registry-server" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.473199 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerName="registry-server" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.473430 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2228f25-b904-46ba-8999-23071a80094c" containerName="mariadb-account-create-update" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.473498 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b5f1b4-0e7f-4b10-808e-845621bc68d2" containerName="mariadb-database-create" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.473577 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf" containerName="mariadb-account-create-update" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.473661 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6360fe51-a8bf-4b95-9d03-163543c7363e" containerName="mariadb-account-create-update" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.473729 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1095391c-df23-4ac0-af2c-b3838d52874d" containerName="ovn-config" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.473786 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0f80a84-53e6-4c55-89a0-fb526f3998c1" containerName="mariadb-database-create" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.473847 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e46c2b-bbc9-46e8-bf17-ead30adb0f2c" containerName="registry-server" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.473957 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cae7741b-9dd1-459a-a05e-c61a3c42b782" containerName="mariadb-database-create" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.474997 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.481862 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.492455 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-tvr8x"] Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.578867 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.578994 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-svc\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.579044 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-config\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.579069 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.579110 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.579131 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrh26\" (UniqueName: \"kubernetes.io/projected/dab30f59-6721-4454-a15a-7163e0766b63-kube-api-access-nrh26\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.614540 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1095391c-df23-4ac0-af2c-b3838d52874d" path="/var/lib/kubelet/pods/1095391c-df23-4ac0-af2c-b3838d52874d/volumes" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.680990 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-svc\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.681101 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-config\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.681163 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.681243 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.681292 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrh26\" (UniqueName: \"kubernetes.io/projected/dab30f59-6721-4454-a15a-7163e0766b63-kube-api-access-nrh26\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.681333 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.682122 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-svc\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.682693 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.682774 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.683153 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-config\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.683238 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.704939 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrh26\" (UniqueName: \"kubernetes.io/projected/dab30f59-6721-4454-a15a-7163e0766b63-kube-api-access-nrh26\") pod \"dnsmasq-dns-764c5664d7-tvr8x\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:41 crc kubenswrapper[4715]: I1210 09:51:41.802109 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.268914 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-tvr8x"] Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.462416 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.601449 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-combined-ca-bundle\") pod \"8ad144d0-9a73-4736-8f88-43dd295478d7\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.601773 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-config-data\") pod \"8ad144d0-9a73-4736-8f88-43dd295478d7\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.601821 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-db-sync-config-data\") pod \"8ad144d0-9a73-4736-8f88-43dd295478d7\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.602011 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2shzm\" (UniqueName: \"kubernetes.io/projected/8ad144d0-9a73-4736-8f88-43dd295478d7-kube-api-access-2shzm\") pod \"8ad144d0-9a73-4736-8f88-43dd295478d7\" (UID: \"8ad144d0-9a73-4736-8f88-43dd295478d7\") " Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.606350 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8ad144d0-9a73-4736-8f88-43dd295478d7" (UID: "8ad144d0-9a73-4736-8f88-43dd295478d7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.608640 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad144d0-9a73-4736-8f88-43dd295478d7-kube-api-access-2shzm" (OuterVolumeSpecName: "kube-api-access-2shzm") pod "8ad144d0-9a73-4736-8f88-43dd295478d7" (UID: "8ad144d0-9a73-4736-8f88-43dd295478d7"). InnerVolumeSpecName "kube-api-access-2shzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.626645 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ad144d0-9a73-4736-8f88-43dd295478d7" (UID: "8ad144d0-9a73-4736-8f88-43dd295478d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.644686 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-config-data" (OuterVolumeSpecName: "config-data") pod "8ad144d0-9a73-4736-8f88-43dd295478d7" (UID: "8ad144d0-9a73-4736-8f88-43dd295478d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.704270 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2shzm\" (UniqueName: \"kubernetes.io/projected/8ad144d0-9a73-4736-8f88-43dd295478d7-kube-api-access-2shzm\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.704543 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.704674 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:42 crc kubenswrapper[4715]: I1210 09:51:42.704771 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ad144d0-9a73-4736-8f88-43dd295478d7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.079944 4715 generic.go:334] "Generic (PLEG): container finished" podID="108964f4-918d-4a16-93d2-85f0ee798efc" containerID="e68e45416c5c46e48d36f198c2ea5aa1081bf6d0dbc2a1e9a18ad82f2c64f29f" exitCode=0 Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.080085 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bhnvd" event={"ID":"108964f4-918d-4a16-93d2-85f0ee798efc","Type":"ContainerDied","Data":"e68e45416c5c46e48d36f198c2ea5aa1081bf6d0dbc2a1e9a18ad82f2c64f29f"} Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.082538 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gd24z" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.082541 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gd24z" event={"ID":"8ad144d0-9a73-4736-8f88-43dd295478d7","Type":"ContainerDied","Data":"1d722e9e1d0bddadb9315d761022d958586c6bc94155e1ba086a5615c1285381"} Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.082586 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d722e9e1d0bddadb9315d761022d958586c6bc94155e1ba086a5615c1285381" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.085001 4715 generic.go:334] "Generic (PLEG): container finished" podID="dab30f59-6721-4454-a15a-7163e0766b63" containerID="17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2" exitCode=0 Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.085047 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" event={"ID":"dab30f59-6721-4454-a15a-7163e0766b63","Type":"ContainerDied","Data":"17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2"} Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.085072 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" event={"ID":"dab30f59-6721-4454-a15a-7163e0766b63","Type":"ContainerStarted","Data":"55400e9e62ce273c205b8ff876bc90719bd6c78f53e6b01bc86a46eb349af703"} Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.458085 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-tvr8x"] Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.493775 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-hh65q"] Dec 10 09:51:43 crc kubenswrapper[4715]: E1210 09:51:43.494237 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad144d0-9a73-4736-8f88-43dd295478d7" containerName="glance-db-sync" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.494265 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad144d0-9a73-4736-8f88-43dd295478d7" containerName="glance-db-sync" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.494498 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad144d0-9a73-4736-8f88-43dd295478d7" containerName="glance-db-sync" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.495639 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.514033 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-hh65q"] Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.619355 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.619432 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.619485 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9rdx\" (UniqueName: \"kubernetes.io/projected/f40297cf-3c69-4709-8139-bb482dbd8c48-kube-api-access-k9rdx\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.619517 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-config\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.619632 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.619697 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.720631 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.720682 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9rdx\" (UniqueName: \"kubernetes.io/projected/f40297cf-3c69-4709-8139-bb482dbd8c48-kube-api-access-k9rdx\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.720704 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-config\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.720742 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.720781 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.720833 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.722101 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-config\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.722128 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.722176 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.722240 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.722276 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.739192 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9rdx\" (UniqueName: \"kubernetes.io/projected/f40297cf-3c69-4709-8139-bb482dbd8c48-kube-api-access-k9rdx\") pod \"dnsmasq-dns-74f6bcbc87-hh65q\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:43 crc kubenswrapper[4715]: I1210 09:51:43.826166 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.107506 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" event={"ID":"dab30f59-6721-4454-a15a-7163e0766b63","Type":"ContainerStarted","Data":"bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3"} Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.107852 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.148401 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" podStartSLOduration=3.148377294 podStartE2EDuration="3.148377294s" podCreationTimestamp="2025-12-10 09:51:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:44.141138949 +0000 UTC m=+1066.884685200" watchObservedRunningTime="2025-12-10 09:51:44.148377294 +0000 UTC m=+1066.891923545" Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.327754 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-hh65q"] Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.473027 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.640278 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-combined-ca-bundle\") pod \"108964f4-918d-4a16-93d2-85f0ee798efc\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.640366 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkl6t\" (UniqueName: \"kubernetes.io/projected/108964f4-918d-4a16-93d2-85f0ee798efc-kube-api-access-gkl6t\") pod \"108964f4-918d-4a16-93d2-85f0ee798efc\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.640460 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-config-data\") pod \"108964f4-918d-4a16-93d2-85f0ee798efc\" (UID: \"108964f4-918d-4a16-93d2-85f0ee798efc\") " Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.645206 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/108964f4-918d-4a16-93d2-85f0ee798efc-kube-api-access-gkl6t" (OuterVolumeSpecName: "kube-api-access-gkl6t") pod "108964f4-918d-4a16-93d2-85f0ee798efc" (UID: "108964f4-918d-4a16-93d2-85f0ee798efc"). InnerVolumeSpecName "kube-api-access-gkl6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.667119 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "108964f4-918d-4a16-93d2-85f0ee798efc" (UID: "108964f4-918d-4a16-93d2-85f0ee798efc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.684600 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-config-data" (OuterVolumeSpecName: "config-data") pod "108964f4-918d-4a16-93d2-85f0ee798efc" (UID: "108964f4-918d-4a16-93d2-85f0ee798efc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.749159 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.749201 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkl6t\" (UniqueName: \"kubernetes.io/projected/108964f4-918d-4a16-93d2-85f0ee798efc-kube-api-access-gkl6t\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:44 crc kubenswrapper[4715]: I1210 09:51:44.749213 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/108964f4-918d-4a16-93d2-85f0ee798efc-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.118493 4715 generic.go:334] "Generic (PLEG): container finished" podID="f40297cf-3c69-4709-8139-bb482dbd8c48" containerID="65fb896967431f44810e9e89107fa3dfead7a2d26315b7beb8789d462a9cfe83" exitCode=0 Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.118575 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" event={"ID":"f40297cf-3c69-4709-8139-bb482dbd8c48","Type":"ContainerDied","Data":"65fb896967431f44810e9e89107fa3dfead7a2d26315b7beb8789d462a9cfe83"} Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.118973 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" event={"ID":"f40297cf-3c69-4709-8139-bb482dbd8c48","Type":"ContainerStarted","Data":"fee77893b763e9751b6009e96b20bb3ead039a50c4db8640ae98201b25a81ff6"} Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.124445 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" podUID="dab30f59-6721-4454-a15a-7163e0766b63" containerName="dnsmasq-dns" containerID="cri-o://bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3" gracePeriod=10 Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.124429 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bhnvd" event={"ID":"108964f4-918d-4a16-93d2-85f0ee798efc","Type":"ContainerDied","Data":"ad30b411c552345ad198331c1400ecd282fe9de56f3f58afe53e9daecc302f8f"} Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.124691 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bhnvd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.124705 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad30b411c552345ad198331c1400ecd282fe9de56f3f58afe53e9daecc302f8f" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.378325 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-hh65q"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.407496 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-25gqd"] Dec 10 09:51:45 crc kubenswrapper[4715]: E1210 09:51:45.407845 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="108964f4-918d-4a16-93d2-85f0ee798efc" containerName="keystone-db-sync" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.407856 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="108964f4-918d-4a16-93d2-85f0ee798efc" containerName="keystone-db-sync" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.408053 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="108964f4-918d-4a16-93d2-85f0ee798efc" containerName="keystone-db-sync" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.408976 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.447403 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-25gqd"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.492991 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-7hv68"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.494209 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.507192 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.507445 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.508119 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-csxkd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.508264 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.508438 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.553357 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7hv68"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.566961 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.567042 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.567081 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmbgj\" (UniqueName: \"kubernetes.io/projected/184e03e0-dcdf-400b-8846-fa2972b4fb42-kube-api-access-mmbgj\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.567184 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-config\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.567219 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.567244 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-svc\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.660997 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2vj6t"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.662454 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.681531 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.681780 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qnst4" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.681949 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.684781 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.684846 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.684887 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmbgj\" (UniqueName: \"kubernetes.io/projected/184e03e0-dcdf-400b-8846-fa2972b4fb42-kube-api-access-mmbgj\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.686038 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-combined-ca-bundle\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.686076 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-scripts\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.686107 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-credential-keys\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.686147 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-fernet-keys\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.686174 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-config\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.686209 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-svc\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.686231 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.686270 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-config-data\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.686295 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65pt5\" (UniqueName: \"kubernetes.io/projected/465326ec-ce56-462a-9c90-ae7294e40b64-kube-api-access-65pt5\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.687390 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.688060 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.689475 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-config\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.690126 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-svc\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.690691 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.702994 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2vj6t"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.727057 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-59f89569c5-knqlx"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.728147 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmbgj\" (UniqueName: \"kubernetes.io/projected/184e03e0-dcdf-400b-8846-fa2972b4fb42-kube-api-access-mmbgj\") pod \"dnsmasq-dns-847c4cc679-25gqd\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.740286 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.760146 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.764974 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-28hsp"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.766345 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.772823 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-2bqf6" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.773209 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.773354 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.773838 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.788165 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-config\") pod \"neutron-db-sync-2vj6t\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.788256 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-combined-ca-bundle\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.788276 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-scripts\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.788294 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-credential-keys\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.788320 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-fernet-keys\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.788352 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-config-data\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.788371 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65pt5\" (UniqueName: \"kubernetes.io/projected/465326ec-ce56-462a-9c90-ae7294e40b64-kube-api-access-65pt5\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.788388 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-combined-ca-bundle\") pod \"neutron-db-sync-2vj6t\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.788467 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn6lm\" (UniqueName: \"kubernetes.io/projected/a08f4204-23b4-425a-a116-cfd37741c87f-kube-api-access-kn6lm\") pod \"neutron-db-sync-2vj6t\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.788557 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-28hsp"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.794262 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.794620 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8dvms" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.794741 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.813737 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-combined-ca-bundle\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.815237 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-config-data\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.817518 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-credential-keys\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.845328 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-scripts\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.854746 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65pt5\" (UniqueName: \"kubernetes.io/projected/465326ec-ce56-462a-9c90-ae7294e40b64-kube-api-access-65pt5\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.854837 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-fernet-keys\") pod \"keystone-bootstrap-7hv68\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.865770 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59f89569c5-knqlx"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.890793 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90e59bef-1c16-4d24-800d-8bac5cb3e990-logs\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.894541 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwm8t\" (UniqueName: \"kubernetes.io/projected/8bf72f3c-8033-4018-b330-72fa2774f402-kube-api-access-dwm8t\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.894617 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn6lm\" (UniqueName: \"kubernetes.io/projected/a08f4204-23b4-425a-a116-cfd37741c87f-kube-api-access-kn6lm\") pod \"neutron-db-sync-2vj6t\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.894705 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-config\") pod \"neutron-db-sync-2vj6t\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.894764 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-db-sync-config-data\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.894791 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-config-data\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.894836 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-combined-ca-bundle\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.894867 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90e59bef-1c16-4d24-800d-8bac5cb3e990-horizon-secret-key\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.894967 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-config-data\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.895052 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bf72f3c-8033-4018-b330-72fa2774f402-etc-machine-id\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.895091 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-scripts\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.895109 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-combined-ca-bundle\") pod \"neutron-db-sync-2vj6t\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.895128 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2jmb\" (UniqueName: \"kubernetes.io/projected/90e59bef-1c16-4d24-800d-8bac5cb3e990-kube-api-access-p2jmb\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.895160 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-scripts\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.912607 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-combined-ca-bundle\") pod \"neutron-db-sync-2vj6t\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.912676 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-zq2xt"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.913737 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.922691 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-config\") pod \"neutron-db-sync-2vj6t\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.923115 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-fnshj" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.923293 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.923400 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.935755 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zq2xt"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.963021 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn6lm\" (UniqueName: \"kubernetes.io/projected/a08f4204-23b4-425a-a116-cfd37741c87f-kube-api-access-kn6lm\") pod \"neutron-db-sync-2vj6t\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.976339 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-gzr48"] Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.977424 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.979817 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9fdrx" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.980027 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 10 09:51:45 crc kubenswrapper[4715]: I1210 09:51:45.982944 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-gzr48"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:45.999813 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwm8t\" (UniqueName: \"kubernetes.io/projected/8bf72f3c-8033-4018-b330-72fa2774f402-kube-api-access-dwm8t\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:45.999937 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-db-sync-config-data\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:45.999979 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-config-data\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.000014 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-combined-ca-bundle\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.000039 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90e59bef-1c16-4d24-800d-8bac5cb3e990-horizon-secret-key\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.000087 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-config-data\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.000166 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bf72f3c-8033-4018-b330-72fa2774f402-etc-machine-id\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.000197 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-scripts\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.000217 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2jmb\" (UniqueName: \"kubernetes.io/projected/90e59bef-1c16-4d24-800d-8bac5cb3e990-kube-api-access-p2jmb\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.000243 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-scripts\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.000272 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90e59bef-1c16-4d24-800d-8bac5cb3e990-logs\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.005776 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90e59bef-1c16-4d24-800d-8bac5cb3e990-logs\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.010670 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90e59bef-1c16-4d24-800d-8bac5cb3e990-horizon-secret-key\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.013781 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-scripts\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.013809 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-config-data\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.013844 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bf72f3c-8033-4018-b330-72fa2774f402-etc-machine-id\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.019290 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-db-sync-config-data\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.022140 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-combined-ca-bundle\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.024763 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.032148 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-config-data\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.032711 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.045098 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-scripts\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.047898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2jmb\" (UniqueName: \"kubernetes.io/projected/90e59bef-1c16-4d24-800d-8bac5cb3e990-kube-api-access-p2jmb\") pod \"horizon-59f89569c5-knqlx\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.052657 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwm8t\" (UniqueName: \"kubernetes.io/projected/8bf72f3c-8033-4018-b330-72fa2774f402-kube-api-access-dwm8t\") pod \"cinder-db-sync-28hsp\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.054106 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-25gqd"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.066672 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6948bcc975-bdnnf"] Dec 10 09:51:46 crc kubenswrapper[4715]: E1210 09:51:46.078439 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab30f59-6721-4454-a15a-7163e0766b63" containerName="init" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.078467 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab30f59-6721-4454-a15a-7163e0766b63" containerName="init" Dec 10 09:51:46 crc kubenswrapper[4715]: E1210 09:51:46.078490 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab30f59-6721-4454-a15a-7163e0766b63" containerName="dnsmasq-dns" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.078496 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab30f59-6721-4454-a15a-7163e0766b63" containerName="dnsmasq-dns" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.089224 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="dab30f59-6721-4454-a15a-7163e0766b63" containerName="dnsmasq-dns" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.098749 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.102692 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-sb\") pod \"dab30f59-6721-4454-a15a-7163e0766b63\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.102774 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-nb\") pod \"dab30f59-6721-4454-a15a-7163e0766b63\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.102859 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-swift-storage-0\") pod \"dab30f59-6721-4454-a15a-7163e0766b63\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.103049 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-config\") pod \"dab30f59-6721-4454-a15a-7163e0766b63\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.103125 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-svc\") pod \"dab30f59-6721-4454-a15a-7163e0766b63\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.103167 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrh26\" (UniqueName: \"kubernetes.io/projected/dab30f59-6721-4454-a15a-7163e0766b63-kube-api-access-nrh26\") pod \"dab30f59-6721-4454-a15a-7163e0766b63\" (UID: \"dab30f59-6721-4454-a15a-7163e0766b63\") " Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.104704 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53507dfd-6f17-486f-8770-c073ec1b42a3-logs\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.104807 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsrcg\" (UniqueName: \"kubernetes.io/projected/53507dfd-6f17-486f-8770-c073ec1b42a3-kube-api-access-wsrcg\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.104874 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-db-sync-config-data\") pod \"barbican-db-sync-gzr48\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.104965 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-scripts\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.105288 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-config-data\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.105368 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-combined-ca-bundle\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.105446 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz8v7\" (UniqueName: \"kubernetes.io/projected/66856339-f4ab-4102-93a8-adc952877ccd-kube-api-access-hz8v7\") pod \"barbican-db-sync-gzr48\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.105565 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-combined-ca-bundle\") pod \"barbican-db-sync-gzr48\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.130334 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.135302 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.141142 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gckbz" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.147552 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.148076 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.143358 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.181668 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dab30f59-6721-4454-a15a-7163e0766b63-kube-api-access-nrh26" (OuterVolumeSpecName: "kube-api-access-nrh26") pod "dab30f59-6721-4454-a15a-7163e0766b63" (UID: "dab30f59-6721-4454-a15a-7163e0766b63"). InnerVolumeSpecName "kube-api-access-nrh26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.196766 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.201028 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" podUID="f40297cf-3c69-4709-8139-bb482dbd8c48" containerName="dnsmasq-dns" containerID="cri-o://9d26e74e207adf9a969aed8826ecf4878574f93f5034e29a4687fb07afe2456f" gracePeriod=10 Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.206610 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-scripts\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.206672 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-logs\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.206702 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53507dfd-6f17-486f-8770-c073ec1b42a3-logs\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.206931 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsrcg\" (UniqueName: \"kubernetes.io/projected/53507dfd-6f17-486f-8770-c073ec1b42a3-kube-api-access-wsrcg\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.206956 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-db-sync-config-data\") pod \"barbican-db-sync-gzr48\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.206971 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be891085-7461-419d-82ae-f01a2643f6b4-logs\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.206991 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-scripts\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207024 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207046 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-config-data\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207064 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhs26\" (UniqueName: \"kubernetes.io/projected/be891085-7461-419d-82ae-f01a2643f6b4-kube-api-access-bhs26\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207099 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-config-data\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207116 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-config-data\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207134 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207153 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-combined-ca-bundle\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207220 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz8v7\" (UniqueName: \"kubernetes.io/projected/66856339-f4ab-4102-93a8-adc952877ccd-kube-api-access-hz8v7\") pod \"barbican-db-sync-gzr48\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207247 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-combined-ca-bundle\") pod \"barbican-db-sync-gzr48\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207266 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-scripts\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207284 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/be891085-7461-419d-82ae-f01a2643f6b4-horizon-secret-key\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207299 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207331 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbzkg\" (UniqueName: \"kubernetes.io/projected/59042857-5b01-491a-826f-4ff16ba96875-kube-api-access-fbzkg\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.207386 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrh26\" (UniqueName: \"kubernetes.io/projected/dab30f59-6721-4454-a15a-7163e0766b63-kube-api-access-nrh26\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.210122 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53507dfd-6f17-486f-8770-c073ec1b42a3-logs\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.211827 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-scripts\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.216388 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" event={"ID":"f40297cf-3c69-4709-8139-bb482dbd8c48","Type":"ContainerStarted","Data":"9d26e74e207adf9a969aed8826ecf4878574f93f5034e29a4687fb07afe2456f"} Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.216439 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.216500 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.217768 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-config-data\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.226980 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.229571 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-combined-ca-bundle\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.238579 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.286346 4715 generic.go:334] "Generic (PLEG): container finished" podID="dab30f59-6721-4454-a15a-7163e0766b63" containerID="bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3" exitCode=0 Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.286401 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" event={"ID":"dab30f59-6721-4454-a15a-7163e0766b63","Type":"ContainerDied","Data":"bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3"} Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.286433 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" event={"ID":"dab30f59-6721-4454-a15a-7163e0766b63","Type":"ContainerDied","Data":"55400e9e62ce273c205b8ff876bc90719bd6c78f53e6b01bc86a46eb349af703"} Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.286455 4715 scope.go:117] "RemoveContainer" containerID="bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.286683 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-tvr8x" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.297246 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.304587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz8v7\" (UniqueName: \"kubernetes.io/projected/66856339-f4ab-4102-93a8-adc952877ccd-kube-api-access-hz8v7\") pod \"barbican-db-sync-gzr48\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.304977 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.306298 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-db-sync-config-data\") pod \"barbican-db-sync-gzr48\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.326741 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-28hsp" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.327808 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-logs\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.327845 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.327890 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be891085-7461-419d-82ae-f01a2643f6b4-logs\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.327934 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-config-data\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.327976 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.328039 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-log-httpd\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.328074 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-config-data\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.328100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhs26\" (UniqueName: \"kubernetes.io/projected/be891085-7461-419d-82ae-f01a2643f6b4-kube-api-access-bhs26\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.328147 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-config-data\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.328707 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-logs\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.329231 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be891085-7461-419d-82ae-f01a2643f6b4-logs\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.330102 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.330170 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-run-httpd\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.330197 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czlpf\" (UniqueName: \"kubernetes.io/projected/2167d98b-0365-421e-9d3d-6ff62f7bfb61-kube-api-access-czlpf\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.330266 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-scripts\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.330301 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/be891085-7461-419d-82ae-f01a2643f6b4-horizon-secret-key\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.330321 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.330356 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbzkg\" (UniqueName: \"kubernetes.io/projected/59042857-5b01-491a-826f-4ff16ba96875-kube-api-access-fbzkg\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.330376 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-scripts\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.330395 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.330418 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-scripts\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.333475 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsrcg\" (UniqueName: \"kubernetes.io/projected/53507dfd-6f17-486f-8770-c073ec1b42a3-kube-api-access-wsrcg\") pod \"placement-db-sync-zq2xt\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.336628 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-scripts\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.337292 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.367583 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-config-data\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.369165 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zq2xt" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.369207 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.370238 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-combined-ca-bundle\") pod \"barbican-db-sync-gzr48\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.377639 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6948bcc975-bdnnf"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.381296 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-config-data\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.391085 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-scripts\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.391481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.391853 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gzr48" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.393321 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhs26\" (UniqueName: \"kubernetes.io/projected/be891085-7461-419d-82ae-f01a2643f6b4-kube-api-access-bhs26\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.412588 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dab30f59-6721-4454-a15a-7163e0766b63" (UID: "dab30f59-6721-4454-a15a-7163e0766b63"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.426388 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/be891085-7461-419d-82ae-f01a2643f6b4-horizon-secret-key\") pod \"horizon-6948bcc975-bdnnf\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.433460 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.434984 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-run-httpd\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.435026 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czlpf\" (UniqueName: \"kubernetes.io/projected/2167d98b-0365-421e-9d3d-6ff62f7bfb61-kube-api-access-czlpf\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.435086 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-scripts\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.435108 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.435167 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.435204 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-config-data\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.435246 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-log-httpd\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.435302 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.435690 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-log-httpd\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.435964 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-run-httpd\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.444762 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.537529 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-scripts\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.558220 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-8jbht"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.559838 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.582212 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-8jbht"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.587139 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.590384 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-config-data\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.594052 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czlpf\" (UniqueName: \"kubernetes.io/projected/2167d98b-0365-421e-9d3d-6ff62f7bfb61-kube-api-access-czlpf\") pod \"ceilometer-0\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.606591 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dab30f59-6721-4454-a15a-7163e0766b63" (UID: "dab30f59-6721-4454-a15a-7163e0766b63"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.606642 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbzkg\" (UniqueName: \"kubernetes.io/projected/59042857-5b01-491a-826f-4ff16ba96875-kube-api-access-fbzkg\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.619780 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.626227 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dab30f59-6721-4454-a15a-7163e0766b63" (UID: "dab30f59-6721-4454-a15a-7163e0766b63"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.644236 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.644274 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.644943 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-config\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.645030 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.645053 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.654184 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.656443 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.659082 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.659222 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkblm\" (UniqueName: \"kubernetes.io/projected/0c33dcac-9cd8-41e9-b4b2-ba521a728286-kube-api-access-gkblm\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.659266 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.661183 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dab30f59-6721-4454-a15a-7163e0766b63" (UID: "dab30f59-6721-4454-a15a-7163e0766b63"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.661247 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.686211 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" podStartSLOduration=3.686192493 podStartE2EDuration="3.686192493s" podCreationTimestamp="2025-12-10 09:51:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:46.354124163 +0000 UTC m=+1069.097670424" watchObservedRunningTime="2025-12-10 09:51:46.686192493 +0000 UTC m=+1069.429738744" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.686663 4715 scope.go:117] "RemoveContainer" containerID="17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.691136 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.691178 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.691194 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.731673 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-config" (OuterVolumeSpecName: "config") pod "dab30f59-6721-4454-a15a-7163e0766b63" (UID: "dab30f59-6721-4454-a15a-7163e0766b63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.735167 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.767330 4715 scope.go:117] "RemoveContainer" containerID="bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3" Dec 10 09:51:46 crc kubenswrapper[4715]: E1210 09:51:46.768100 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3\": container with ID starting with bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3 not found: ID does not exist" containerID="bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.768129 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3"} err="failed to get container status \"bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3\": rpc error: code = NotFound desc = could not find container \"bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3\": container with ID starting with bf613e4eb245da107f0c503d562850ad736fc8940ba000a22f58e449875908f3 not found: ID does not exist" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.768151 4715 scope.go:117] "RemoveContainer" containerID="17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2" Dec 10 09:51:46 crc kubenswrapper[4715]: E1210 09:51:46.768689 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2\": container with ID starting with 17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2 not found: ID does not exist" containerID="17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.768708 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2"} err="failed to get container status \"17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2\": rpc error: code = NotFound desc = could not find container \"17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2\": container with ID starting with 17371b3144ee257b5eed4448a002a4668c3ff620b6dc8b7bbc0f63b2fa7ea6a2 not found: ID does not exist" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.792808 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw2wn\" (UniqueName: \"kubernetes.io/projected/b8e3652a-3b12-41d2-983e-83e49d6df137-kube-api-access-bw2wn\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.792887 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-config\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.792945 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.792968 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.793031 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.793066 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-logs\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.793100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.793150 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.793182 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkblm\" (UniqueName: \"kubernetes.io/projected/0c33dcac-9cd8-41e9-b4b2-ba521a728286-kube-api-access-gkblm\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.793506 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.793540 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.793626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.793679 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.793767 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab30f59-6721-4454-a15a-7163e0766b63-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.795560 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.795627 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-25gqd"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.795844 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.796416 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.796614 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-config\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.796625 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.806416 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.830305 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkblm\" (UniqueName: \"kubernetes.io/projected/0c33dcac-9cd8-41e9-b4b2-ba521a728286-kube-api-access-gkblm\") pod \"dnsmasq-dns-785d8bcb8c-8jbht\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.897379 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw2wn\" (UniqueName: \"kubernetes.io/projected/b8e3652a-3b12-41d2-983e-83e49d6df137-kube-api-access-bw2wn\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.898512 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.898657 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.899005 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-logs\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.899259 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.900470 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.901669 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.901823 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.902287 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.901707 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-logs\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.910905 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.914386 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.915474 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.945824 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw2wn\" (UniqueName: \"kubernetes.io/projected/b8e3652a-3b12-41d2-983e-83e49d6df137-kube-api-access-bw2wn\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.951256 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.972509 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2vj6t"] Dec 10 09:51:46 crc kubenswrapper[4715]: I1210 09:51:46.997039 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.007019 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.209056 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-tvr8x"] Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.240537 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-tvr8x"] Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.366546 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7hv68"] Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.379336 4715 generic.go:334] "Generic (PLEG): container finished" podID="f40297cf-3c69-4709-8139-bb482dbd8c48" containerID="9d26e74e207adf9a969aed8826ecf4878574f93f5034e29a4687fb07afe2456f" exitCode=0 Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.379443 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" event={"ID":"f40297cf-3c69-4709-8139-bb482dbd8c48","Type":"ContainerDied","Data":"9d26e74e207adf9a969aed8826ecf4878574f93f5034e29a4687fb07afe2456f"} Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.394580 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2vj6t" event={"ID":"a08f4204-23b4-425a-a116-cfd37741c87f","Type":"ContainerStarted","Data":"ea9785ad784790c9925ffbeee6c2aa38af92552cd0171c9f66388b1cac862126"} Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.411285 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-25gqd" event={"ID":"184e03e0-dcdf-400b-8846-fa2972b4fb42","Type":"ContainerStarted","Data":"252f971de7420610fcf6ef3ec3067c9a83a0e0442c43f7400bcb83a759e42dde"} Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.446073 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.519278 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59f89569c5-knqlx"] Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.536888 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zq2xt"] Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.561601 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-swift-storage-0\") pod \"f40297cf-3c69-4709-8139-bb482dbd8c48\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.561806 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9rdx\" (UniqueName: \"kubernetes.io/projected/f40297cf-3c69-4709-8139-bb482dbd8c48-kube-api-access-k9rdx\") pod \"f40297cf-3c69-4709-8139-bb482dbd8c48\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.561865 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-sb\") pod \"f40297cf-3c69-4709-8139-bb482dbd8c48\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.561931 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-config\") pod \"f40297cf-3c69-4709-8139-bb482dbd8c48\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.561962 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-svc\") pod \"f40297cf-3c69-4709-8139-bb482dbd8c48\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.561984 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-nb\") pod \"f40297cf-3c69-4709-8139-bb482dbd8c48\" (UID: \"f40297cf-3c69-4709-8139-bb482dbd8c48\") " Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.641865 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f40297cf-3c69-4709-8139-bb482dbd8c48" (UID: "f40297cf-3c69-4709-8139-bb482dbd8c48"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.644636 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f40297cf-3c69-4709-8139-bb482dbd8c48-kube-api-access-k9rdx" (OuterVolumeSpecName: "kube-api-access-k9rdx") pod "f40297cf-3c69-4709-8139-bb482dbd8c48" (UID: "f40297cf-3c69-4709-8139-bb482dbd8c48"). InnerVolumeSpecName "kube-api-access-k9rdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.657224 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dab30f59-6721-4454-a15a-7163e0766b63" path="/var/lib/kubelet/pods/dab30f59-6721-4454-a15a-7163e0766b63/volumes" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.657699 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f40297cf-3c69-4709-8139-bb482dbd8c48" (UID: "f40297cf-3c69-4709-8139-bb482dbd8c48"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.664927 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.666996 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9rdx\" (UniqueName: \"kubernetes.io/projected/f40297cf-3c69-4709-8139-bb482dbd8c48-kube-api-access-k9rdx\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.667013 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.695161 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-config" (OuterVolumeSpecName: "config") pod "f40297cf-3c69-4709-8139-bb482dbd8c48" (UID: "f40297cf-3c69-4709-8139-bb482dbd8c48"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.719495 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f40297cf-3c69-4709-8139-bb482dbd8c48" (UID: "f40297cf-3c69-4709-8139-bb482dbd8c48"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.749294 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f40297cf-3c69-4709-8139-bb482dbd8c48" (UID: "f40297cf-3c69-4709-8139-bb482dbd8c48"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.773216 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.773252 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:47 crc kubenswrapper[4715]: I1210 09:51:47.773262 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f40297cf-3c69-4709-8139-bb482dbd8c48-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.018621 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-28hsp"] Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.036671 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-gzr48"] Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.046859 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6948bcc975-bdnnf"] Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.055150 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:51:48 crc kubenswrapper[4715]: W1210 09:51:48.061777 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66856339_f4ab_4102_93a8_adc952877ccd.slice/crio-cb728b42ad1d452c5e1bc5c830dc1a5d39e4abe4cca259bed26503c9eeed0eec WatchSource:0}: Error finding container cb728b42ad1d452c5e1bc5c830dc1a5d39e4abe4cca259bed26503c9eeed0eec: Status 404 returned error can't find the container with id cb728b42ad1d452c5e1bc5c830dc1a5d39e4abe4cca259bed26503c9eeed0eec Dec 10 09:51:48 crc kubenswrapper[4715]: W1210 09:51:48.064613 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2167d98b_0365_421e_9d3d_6ff62f7bfb61.slice/crio-d48c3f8c7911397fdca54dde3b2bac53d0fb6a4cd755e50728d3519f58e22c52 WatchSource:0}: Error finding container d48c3f8c7911397fdca54dde3b2bac53d0fb6a4cd755e50728d3519f58e22c52: Status 404 returned error can't find the container with id d48c3f8c7911397fdca54dde3b2bac53d0fb6a4cd755e50728d3519f58e22c52 Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.162944 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-8jbht"] Dec 10 09:51:48 crc kubenswrapper[4715]: W1210 09:51:48.179418 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c33dcac_9cd8_41e9_b4b2_ba521a728286.slice/crio-63c22a057e78cb913772e8849661d111d991d6f7593f2b1ca61f99f4b9e51b87 WatchSource:0}: Error finding container 63c22a057e78cb913772e8849661d111d991d6f7593f2b1ca61f99f4b9e51b87: Status 404 returned error can't find the container with id 63c22a057e78cb913772e8849661d111d991d6f7593f2b1ca61f99f4b9e51b87 Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.260592 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:51:48 crc kubenswrapper[4715]: W1210 09:51:48.265645 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59042857_5b01_491a_826f_4ff16ba96875.slice/crio-9edb53a5044a9dd6f428feb281b03e446407169c07f940b98b99de53b66cc46e WatchSource:0}: Error finding container 9edb53a5044a9dd6f428feb281b03e446407169c07f940b98b99de53b66cc46e: Status 404 returned error can't find the container with id 9edb53a5044a9dd6f428feb281b03e446407169c07f940b98b99de53b66cc46e Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.425878 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"59042857-5b01-491a-826f-4ff16ba96875","Type":"ContainerStarted","Data":"9edb53a5044a9dd6f428feb281b03e446407169c07f940b98b99de53b66cc46e"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.428171 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gzr48" event={"ID":"66856339-f4ab-4102-93a8-adc952877ccd","Type":"ContainerStarted","Data":"cb728b42ad1d452c5e1bc5c830dc1a5d39e4abe4cca259bed26503c9eeed0eec"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.430000 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" event={"ID":"f40297cf-3c69-4709-8139-bb482dbd8c48","Type":"ContainerDied","Data":"fee77893b763e9751b6009e96b20bb3ead039a50c4db8640ae98201b25a81ff6"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.430046 4715 scope.go:117] "RemoveContainer" containerID="9d26e74e207adf9a969aed8826ecf4878574f93f5034e29a4687fb07afe2456f" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.430172 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-hh65q" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.446803 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2167d98b-0365-421e-9d3d-6ff62f7bfb61","Type":"ContainerStarted","Data":"d48c3f8c7911397fdca54dde3b2bac53d0fb6a4cd755e50728d3519f58e22c52"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.449305 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7hv68" event={"ID":"465326ec-ce56-462a-9c90-ae7294e40b64","Type":"ContainerStarted","Data":"e3c748c38073b710935f8bc8f46b6ae6913e36215588881f9fdd6733d13a7b41"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.449329 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7hv68" event={"ID":"465326ec-ce56-462a-9c90-ae7294e40b64","Type":"ContainerStarted","Data":"2ef50d81e4152f179b7219e03c228c0d308f2fb3a9ffba5c0c51d27b2e2b4996"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.453475 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2vj6t" event={"ID":"a08f4204-23b4-425a-a116-cfd37741c87f","Type":"ContainerStarted","Data":"0c4f3f531995496685f907c8d2b13f8583be7b428dd7c227e2622600162bb714"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.460429 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.503228 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59f89569c5-knqlx" event={"ID":"90e59bef-1c16-4d24-800d-8bac5cb3e990","Type":"ContainerStarted","Data":"c161b0e868272e83c4ee1c3ca7ed7f7829257b735b3c5861020bb79d41b71a6a"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.504876 4715 scope.go:117] "RemoveContainer" containerID="65fb896967431f44810e9e89107fa3dfead7a2d26315b7beb8789d462a9cfe83" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.517504 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6948bcc975-bdnnf" event={"ID":"be891085-7461-419d-82ae-f01a2643f6b4","Type":"ContainerStarted","Data":"14915d715ef83133cb338df20bed8f993cf984ebded8b0ce6deae033f0944536"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.525468 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-7hv68" podStartSLOduration=3.525441776 podStartE2EDuration="3.525441776s" podCreationTimestamp="2025-12-10 09:51:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:48.473069944 +0000 UTC m=+1071.216616195" watchObservedRunningTime="2025-12-10 09:51:48.525441776 +0000 UTC m=+1071.268988027" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.547835 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2vj6t" podStartSLOduration=3.547813479 podStartE2EDuration="3.547813479s" podCreationTimestamp="2025-12-10 09:51:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:48.496731084 +0000 UTC m=+1071.240277335" watchObservedRunningTime="2025-12-10 09:51:48.547813479 +0000 UTC m=+1071.291359740" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.578842 4715 generic.go:334] "Generic (PLEG): container finished" podID="184e03e0-dcdf-400b-8846-fa2972b4fb42" containerID="b584a9264fb176c73ccc2261b9800635bf055839b2884ea8c3c51d2ee3899572" exitCode=0 Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.578901 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-25gqd" event={"ID":"184e03e0-dcdf-400b-8846-fa2972b4fb42","Type":"ContainerDied","Data":"b584a9264fb176c73ccc2261b9800635bf055839b2884ea8c3c51d2ee3899572"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.625586 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-hh65q"] Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.628727 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" event={"ID":"0c33dcac-9cd8-41e9-b4b2-ba521a728286","Type":"ContainerStarted","Data":"63c22a057e78cb913772e8849661d111d991d6f7593f2b1ca61f99f4b9e51b87"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.668107 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-hh65q"] Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.672388 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zq2xt" event={"ID":"53507dfd-6f17-486f-8770-c073ec1b42a3","Type":"ContainerStarted","Data":"2f3fc708d59ff69d9105a8e945f2864324b8786ef2108b561cd706bb6ecdfb86"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.691672 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.695946 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-28hsp" event={"ID":"8bf72f3c-8033-4018-b330-72fa2774f402","Type":"ContainerStarted","Data":"cd889dd567a1b1de84aff2de0dcc0debb5051bf39cfffbdfb6abd287785cc536"} Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.878260 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59f89569c5-knqlx"] Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.889009 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7fcc456ff5-rc25b"] Dec 10 09:51:48 crc kubenswrapper[4715]: E1210 09:51:48.889365 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40297cf-3c69-4709-8139-bb482dbd8c48" containerName="init" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.889376 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40297cf-3c69-4709-8139-bb482dbd8c48" containerName="init" Dec 10 09:51:48 crc kubenswrapper[4715]: E1210 09:51:48.889408 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40297cf-3c69-4709-8139-bb482dbd8c48" containerName="dnsmasq-dns" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.889414 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40297cf-3c69-4709-8139-bb482dbd8c48" containerName="dnsmasq-dns" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.889584 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40297cf-3c69-4709-8139-bb482dbd8c48" containerName="dnsmasq-dns" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.890541 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.909987 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.928020 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fcc456ff5-rc25b"] Dec 10 09:51:48 crc kubenswrapper[4715]: I1210 09:51:48.939961 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.054351 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-config-data\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.054408 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5479ff20-2e76-4be0-807b-eca95c4cea9b-logs\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.054452 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5479ff20-2e76-4be0-807b-eca95c4cea9b-horizon-secret-key\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.054492 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-scripts\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.054520 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s894m\" (UniqueName: \"kubernetes.io/projected/5479ff20-2e76-4be0-807b-eca95c4cea9b-kube-api-access-s894m\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.155649 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-config-data\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.155704 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5479ff20-2e76-4be0-807b-eca95c4cea9b-logs\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.155740 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5479ff20-2e76-4be0-807b-eca95c4cea9b-horizon-secret-key\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.155781 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-scripts\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.155814 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s894m\" (UniqueName: \"kubernetes.io/projected/5479ff20-2e76-4be0-807b-eca95c4cea9b-kube-api-access-s894m\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.157730 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-config-data\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.158178 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-scripts\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.158406 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5479ff20-2e76-4be0-807b-eca95c4cea9b-logs\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.178604 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5479ff20-2e76-4be0-807b-eca95c4cea9b-horizon-secret-key\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.190460 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s894m\" (UniqueName: \"kubernetes.io/projected/5479ff20-2e76-4be0-807b-eca95c4cea9b-kube-api-access-s894m\") pod \"horizon-7fcc456ff5-rc25b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.248412 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:51:49 crc kubenswrapper[4715]: I1210 09:51:49.455587 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.568667 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-svc\") pod \"184e03e0-dcdf-400b-8846-fa2972b4fb42\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.568813 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-nb\") pod \"184e03e0-dcdf-400b-8846-fa2972b4fb42\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.568840 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-sb\") pod \"184e03e0-dcdf-400b-8846-fa2972b4fb42\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.569192 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-swift-storage-0\") pod \"184e03e0-dcdf-400b-8846-fa2972b4fb42\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.569272 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmbgj\" (UniqueName: \"kubernetes.io/projected/184e03e0-dcdf-400b-8846-fa2972b4fb42-kube-api-access-mmbgj\") pod \"184e03e0-dcdf-400b-8846-fa2972b4fb42\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.569311 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-config\") pod \"184e03e0-dcdf-400b-8846-fa2972b4fb42\" (UID: \"184e03e0-dcdf-400b-8846-fa2972b4fb42\") " Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.576561 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/184e03e0-dcdf-400b-8846-fa2972b4fb42-kube-api-access-mmbgj" (OuterVolumeSpecName: "kube-api-access-mmbgj") pod "184e03e0-dcdf-400b-8846-fa2972b4fb42" (UID: "184e03e0-dcdf-400b-8846-fa2972b4fb42"). InnerVolumeSpecName "kube-api-access-mmbgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.606610 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "184e03e0-dcdf-400b-8846-fa2972b4fb42" (UID: "184e03e0-dcdf-400b-8846-fa2972b4fb42"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.607946 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "184e03e0-dcdf-400b-8846-fa2972b4fb42" (UID: "184e03e0-dcdf-400b-8846-fa2972b4fb42"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.628220 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "184e03e0-dcdf-400b-8846-fa2972b4fb42" (UID: "184e03e0-dcdf-400b-8846-fa2972b4fb42"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.630748 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-config" (OuterVolumeSpecName: "config") pod "184e03e0-dcdf-400b-8846-fa2972b4fb42" (UID: "184e03e0-dcdf-400b-8846-fa2972b4fb42"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.636696 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f40297cf-3c69-4709-8139-bb482dbd8c48" path="/var/lib/kubelet/pods/f40297cf-3c69-4709-8139-bb482dbd8c48/volumes" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.641022 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "184e03e0-dcdf-400b-8846-fa2972b4fb42" (UID: "184e03e0-dcdf-400b-8846-fa2972b4fb42"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.673403 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmbgj\" (UniqueName: \"kubernetes.io/projected/184e03e0-dcdf-400b-8846-fa2972b4fb42-kube-api-access-mmbgj\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.673441 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.673452 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.673463 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.673473 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.673483 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/184e03e0-dcdf-400b-8846-fa2972b4fb42-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.721523 4715 generic.go:334] "Generic (PLEG): container finished" podID="0c33dcac-9cd8-41e9-b4b2-ba521a728286" containerID="9abec03902435ff97b9dc6dfc48747fe8fb38d18f71a7cb6cccbd9c7b493001c" exitCode=0 Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.721621 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" event={"ID":"0c33dcac-9cd8-41e9-b4b2-ba521a728286","Type":"ContainerDied","Data":"9abec03902435ff97b9dc6dfc48747fe8fb38d18f71a7cb6cccbd9c7b493001c"} Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.733949 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-25gqd" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.733966 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-25gqd" event={"ID":"184e03e0-dcdf-400b-8846-fa2972b4fb42","Type":"ContainerDied","Data":"252f971de7420610fcf6ef3ec3067c9a83a0e0442c43f7400bcb83a759e42dde"} Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.734049 4715 scope.go:117] "RemoveContainer" containerID="b584a9264fb176c73ccc2261b9800635bf055839b2884ea8c3c51d2ee3899572" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:49.774534 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b8e3652a-3b12-41d2-983e-83e49d6df137","Type":"ContainerStarted","Data":"e676518c27c172350ddcbd19f95ff35ab0d2b23010375c050b24cff484e59e2b"} Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:50.084653 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-25gqd"] Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:50.102762 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-25gqd"] Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:50.803576 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"59042857-5b01-491a-826f-4ff16ba96875","Type":"ContainerStarted","Data":"91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461"} Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:50.822358 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b8e3652a-3b12-41d2-983e-83e49d6df137","Type":"ContainerStarted","Data":"63db12bc74a53c2de57ce51c452d6033c15a94bc92a94f9eb8aa4e6a25903b05"} Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:50.825438 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" event={"ID":"0c33dcac-9cd8-41e9-b4b2-ba521a728286","Type":"ContainerStarted","Data":"8cabc5a59d136bdb08c1067fda57027192f115e00ce85b89dd35bcc8c305f28c"} Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:50.825821 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:50 crc kubenswrapper[4715]: I1210 09:51:50.850427 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" podStartSLOduration=5.850407211 podStartE2EDuration="5.850407211s" podCreationTimestamp="2025-12-10 09:51:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:50.845177643 +0000 UTC m=+1073.588723904" watchObservedRunningTime="2025-12-10 09:51:50.850407211 +0000 UTC m=+1073.593953462" Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.168564 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fcc456ff5-rc25b"] Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.629206 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="184e03e0-dcdf-400b-8846-fa2972b4fb42" path="/var/lib/kubelet/pods/184e03e0-dcdf-400b-8846-fa2972b4fb42/volumes" Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.860696 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fcc456ff5-rc25b" event={"ID":"5479ff20-2e76-4be0-807b-eca95c4cea9b","Type":"ContainerStarted","Data":"a2a1703f47703be70719d2cb495de52284213e1863097341da7a601973029779"} Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.873091 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"59042857-5b01-491a-826f-4ff16ba96875","Type":"ContainerStarted","Data":"0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043"} Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.873193 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="59042857-5b01-491a-826f-4ff16ba96875" containerName="glance-httpd" containerID="cri-o://0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043" gracePeriod=30 Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.873175 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="59042857-5b01-491a-826f-4ff16ba96875" containerName="glance-log" containerID="cri-o://91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461" gracePeriod=30 Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.878644 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b8e3652a-3b12-41d2-983e-83e49d6df137","Type":"ContainerStarted","Data":"c5aeac5a1f59add4398e69a067c7f3206bbe2c372ba9ca2753e784a5728eeb51"} Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.878826 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b8e3652a-3b12-41d2-983e-83e49d6df137" containerName="glance-httpd" containerID="cri-o://c5aeac5a1f59add4398e69a067c7f3206bbe2c372ba9ca2753e784a5728eeb51" gracePeriod=30 Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.878807 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b8e3652a-3b12-41d2-983e-83e49d6df137" containerName="glance-log" containerID="cri-o://63db12bc74a53c2de57ce51c452d6033c15a94bc92a94f9eb8aa4e6a25903b05" gracePeriod=30 Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.921755 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.9217326759999995 podStartE2EDuration="6.921732676s" podCreationTimestamp="2025-12-10 09:51:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:51.89464631 +0000 UTC m=+1074.638192561" watchObservedRunningTime="2025-12-10 09:51:51.921732676 +0000 UTC m=+1074.665278927" Dec 10 09:51:51 crc kubenswrapper[4715]: I1210 09:51:51.927466 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.927449988 podStartE2EDuration="5.927449988s" podCreationTimestamp="2025-12-10 09:51:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:51:51.916681913 +0000 UTC m=+1074.660228164" watchObservedRunningTime="2025-12-10 09:51:51.927449988 +0000 UTC m=+1074.670996239" Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.786174 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.919320 4715 generic.go:334] "Generic (PLEG): container finished" podID="b8e3652a-3b12-41d2-983e-83e49d6df137" containerID="c5aeac5a1f59add4398e69a067c7f3206bbe2c372ba9ca2753e784a5728eeb51" exitCode=0 Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.919575 4715 generic.go:334] "Generic (PLEG): container finished" podID="b8e3652a-3b12-41d2-983e-83e49d6df137" containerID="63db12bc74a53c2de57ce51c452d6033c15a94bc92a94f9eb8aa4e6a25903b05" exitCode=143 Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.919642 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b8e3652a-3b12-41d2-983e-83e49d6df137","Type":"ContainerDied","Data":"c5aeac5a1f59add4398e69a067c7f3206bbe2c372ba9ca2753e784a5728eeb51"} Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.919669 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b8e3652a-3b12-41d2-983e-83e49d6df137","Type":"ContainerDied","Data":"63db12bc74a53c2de57ce51c452d6033c15a94bc92a94f9eb8aa4e6a25903b05"} Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.926317 4715 generic.go:334] "Generic (PLEG): container finished" podID="465326ec-ce56-462a-9c90-ae7294e40b64" containerID="e3c748c38073b710935f8bc8f46b6ae6913e36215588881f9fdd6733d13a7b41" exitCode=0 Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.926376 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7hv68" event={"ID":"465326ec-ce56-462a-9c90-ae7294e40b64","Type":"ContainerDied","Data":"e3c748c38073b710935f8bc8f46b6ae6913e36215588881f9fdd6733d13a7b41"} Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.935357 4715 generic.go:334] "Generic (PLEG): container finished" podID="59042857-5b01-491a-826f-4ff16ba96875" containerID="0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043" exitCode=0 Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.935383 4715 generic.go:334] "Generic (PLEG): container finished" podID="59042857-5b01-491a-826f-4ff16ba96875" containerID="91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461" exitCode=143 Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.935404 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"59042857-5b01-491a-826f-4ff16ba96875","Type":"ContainerDied","Data":"0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043"} Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.935433 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"59042857-5b01-491a-826f-4ff16ba96875","Type":"ContainerDied","Data":"91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461"} Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.935443 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"59042857-5b01-491a-826f-4ff16ba96875","Type":"ContainerDied","Data":"9edb53a5044a9dd6f428feb281b03e446407169c07f940b98b99de53b66cc46e"} Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.935461 4715 scope.go:117] "RemoveContainer" containerID="0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043" Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.935455 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.954821 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbzkg\" (UniqueName: \"kubernetes.io/projected/59042857-5b01-491a-826f-4ff16ba96875-kube-api-access-fbzkg\") pod \"59042857-5b01-491a-826f-4ff16ba96875\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.954853 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"59042857-5b01-491a-826f-4ff16ba96875\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.954980 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-config-data\") pod \"59042857-5b01-491a-826f-4ff16ba96875\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.955013 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-logs\") pod \"59042857-5b01-491a-826f-4ff16ba96875\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.955051 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-scripts\") pod \"59042857-5b01-491a-826f-4ff16ba96875\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.955102 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-httpd-run\") pod \"59042857-5b01-491a-826f-4ff16ba96875\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.955118 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-combined-ca-bundle\") pod \"59042857-5b01-491a-826f-4ff16ba96875\" (UID: \"59042857-5b01-491a-826f-4ff16ba96875\") " Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.961453 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "59042857-5b01-491a-826f-4ff16ba96875" (UID: "59042857-5b01-491a-826f-4ff16ba96875"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.961598 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-logs" (OuterVolumeSpecName: "logs") pod "59042857-5b01-491a-826f-4ff16ba96875" (UID: "59042857-5b01-491a-826f-4ff16ba96875"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.970229 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-scripts" (OuterVolumeSpecName: "scripts") pod "59042857-5b01-491a-826f-4ff16ba96875" (UID: "59042857-5b01-491a-826f-4ff16ba96875"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.970260 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "59042857-5b01-491a-826f-4ff16ba96875" (UID: "59042857-5b01-491a-826f-4ff16ba96875"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.970297 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59042857-5b01-491a-826f-4ff16ba96875-kube-api-access-fbzkg" (OuterVolumeSpecName: "kube-api-access-fbzkg") pod "59042857-5b01-491a-826f-4ff16ba96875" (UID: "59042857-5b01-491a-826f-4ff16ba96875"). InnerVolumeSpecName "kube-api-access-fbzkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:51:52 crc kubenswrapper[4715]: I1210 09:51:52.997406 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59042857-5b01-491a-826f-4ff16ba96875" (UID: "59042857-5b01-491a-826f-4ff16ba96875"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.013116 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-config-data" (OuterVolumeSpecName: "config-data") pod "59042857-5b01-491a-826f-4ff16ba96875" (UID: "59042857-5b01-491a-826f-4ff16ba96875"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.061520 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.061597 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.061612 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.061622 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59042857-5b01-491a-826f-4ff16ba96875-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.061634 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59042857-5b01-491a-826f-4ff16ba96875-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.061649 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbzkg\" (UniqueName: \"kubernetes.io/projected/59042857-5b01-491a-826f-4ff16ba96875-kube-api-access-fbzkg\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.061694 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.097877 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.164105 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.277931 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.303051 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.311460 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:51:53 crc kubenswrapper[4715]: E1210 09:51:53.311958 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59042857-5b01-491a-826f-4ff16ba96875" containerName="glance-log" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.311979 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="59042857-5b01-491a-826f-4ff16ba96875" containerName="glance-log" Dec 10 09:51:53 crc kubenswrapper[4715]: E1210 09:51:53.312003 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59042857-5b01-491a-826f-4ff16ba96875" containerName="glance-httpd" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.312009 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="59042857-5b01-491a-826f-4ff16ba96875" containerName="glance-httpd" Dec 10 09:51:53 crc kubenswrapper[4715]: E1210 09:51:53.312027 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="184e03e0-dcdf-400b-8846-fa2972b4fb42" containerName="init" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.312033 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="184e03e0-dcdf-400b-8846-fa2972b4fb42" containerName="init" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.312197 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="59042857-5b01-491a-826f-4ff16ba96875" containerName="glance-httpd" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.312217 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="59042857-5b01-491a-826f-4ff16ba96875" containerName="glance-log" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.312231 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="184e03e0-dcdf-400b-8846-fa2972b4fb42" containerName="init" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.313156 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.316463 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.321195 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.479391 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-config-data\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.479501 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.479545 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-scripts\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.479582 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq4dg\" (UniqueName: \"kubernetes.io/projected/7c0209db-d227-4d43-a132-e2393ce0d5c3-kube-api-access-wq4dg\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.479769 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.479811 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-logs\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.480128 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.582577 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.582762 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-config-data\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.582810 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.582847 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-scripts\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.582891 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq4dg\" (UniqueName: \"kubernetes.io/projected/7c0209db-d227-4d43-a132-e2393ce0d5c3-kube-api-access-wq4dg\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.582943 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-logs\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.582964 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.582966 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.589759 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.590391 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.592313 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-logs\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.595891 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-scripts\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.604116 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq4dg\" (UniqueName: \"kubernetes.io/projected/7c0209db-d227-4d43-a132-e2393ce0d5c3-kube-api-access-wq4dg\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.613741 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-config-data\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.619586 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.649716 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.665516 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59042857-5b01-491a-826f-4ff16ba96875" path="/var/lib/kubelet/pods/59042857-5b01-491a-826f-4ff16ba96875/volumes" Dec 10 09:51:53 crc kubenswrapper[4715]: I1210 09:51:53.965670 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.494727 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6948bcc975-bdnnf"] Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.526389 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67f6cc7984-tshct"] Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.530966 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.533536 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.544570 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67f6cc7984-tshct"] Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.557759 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-combined-ca-bundle\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.557811 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-tls-certs\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.557859 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-config-data\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.557949 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7czs\" (UniqueName: \"kubernetes.io/projected/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-kube-api-access-n7czs\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.558012 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-logs\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.558037 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-secret-key\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.558075 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-scripts\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.655530 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fcc456ff5-rc25b"] Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.660319 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-logs\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.660411 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-secret-key\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.660453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-scripts\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.660559 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-combined-ca-bundle\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.660579 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-tls-certs\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.660608 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-config-data\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.660668 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7czs\" (UniqueName: \"kubernetes.io/projected/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-kube-api-access-n7czs\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.660823 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-logs\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.661809 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-scripts\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.662505 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-config-data\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.666954 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-tls-certs\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.667279 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-combined-ca-bundle\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.684665 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-secret-key\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.702307 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7czs\" (UniqueName: \"kubernetes.io/projected/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-kube-api-access-n7czs\") pod \"horizon-67f6cc7984-tshct\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.710440 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-857b4dfcd4-wnlbz"] Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.712158 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.719283 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-857b4dfcd4-wnlbz"] Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.763116 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4214260f-423f-47fb-b062-c752f3519175-horizon-secret-key\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.763244 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4214260f-423f-47fb-b062-c752f3519175-combined-ca-bundle\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.763362 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzn2s\" (UniqueName: \"kubernetes.io/projected/4214260f-423f-47fb-b062-c752f3519175-kube-api-access-lzn2s\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.763420 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4214260f-423f-47fb-b062-c752f3519175-horizon-tls-certs\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.763662 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4214260f-423f-47fb-b062-c752f3519175-scripts\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.763750 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4214260f-423f-47fb-b062-c752f3519175-logs\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.763848 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4214260f-423f-47fb-b062-c752f3519175-config-data\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.859640 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.866045 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4214260f-423f-47fb-b062-c752f3519175-combined-ca-bundle\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.866124 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzn2s\" (UniqueName: \"kubernetes.io/projected/4214260f-423f-47fb-b062-c752f3519175-kube-api-access-lzn2s\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.866158 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4214260f-423f-47fb-b062-c752f3519175-horizon-tls-certs\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.866205 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4214260f-423f-47fb-b062-c752f3519175-scripts\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.866252 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4214260f-423f-47fb-b062-c752f3519175-logs\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.866299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4214260f-423f-47fb-b062-c752f3519175-config-data\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.866351 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4214260f-423f-47fb-b062-c752f3519175-horizon-secret-key\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.868381 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4214260f-423f-47fb-b062-c752f3519175-logs\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.868727 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4214260f-423f-47fb-b062-c752f3519175-scripts\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.870544 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4214260f-423f-47fb-b062-c752f3519175-config-data\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.872132 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4214260f-423f-47fb-b062-c752f3519175-combined-ca-bundle\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.873314 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4214260f-423f-47fb-b062-c752f3519175-horizon-tls-certs\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.890291 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4214260f-423f-47fb-b062-c752f3519175-horizon-secret-key\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.902694 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzn2s\" (UniqueName: \"kubernetes.io/projected/4214260f-423f-47fb-b062-c752f3519175-kube-api-access-lzn2s\") pod \"horizon-857b4dfcd4-wnlbz\" (UID: \"4214260f-423f-47fb-b062-c752f3519175\") " pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:56 crc kubenswrapper[4715]: I1210 09:51:56.952149 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:51:57 crc kubenswrapper[4715]: I1210 09:51:57.049516 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-d462h"] Dec 10 09:51:57 crc kubenswrapper[4715]: I1210 09:51:57.049842 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-d462h" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerName="dnsmasq-dns" containerID="cri-o://82c73dd9e350a16ff73aac255bc07d2d8de274b87dacef3429c9d880d308c897" gracePeriod=10 Dec 10 09:51:57 crc kubenswrapper[4715]: I1210 09:51:57.081579 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:51:58 crc kubenswrapper[4715]: I1210 09:51:58.510628 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-d462h" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Dec 10 09:51:59 crc kubenswrapper[4715]: I1210 09:51:59.008933 4715 generic.go:334] "Generic (PLEG): container finished" podID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerID="82c73dd9e350a16ff73aac255bc07d2d8de274b87dacef3429c9d880d308c897" exitCode=0 Dec 10 09:51:59 crc kubenswrapper[4715]: I1210 09:51:59.009012 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-d462h" event={"ID":"de02f9b6-491d-4be4-9f06-7c1c53976694","Type":"ContainerDied","Data":"82c73dd9e350a16ff73aac255bc07d2d8de274b87dacef3429c9d880d308c897"} Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.927303 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.986536 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-scripts\") pod \"b8e3652a-3b12-41d2-983e-83e49d6df137\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.986604 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-combined-ca-bundle\") pod \"b8e3652a-3b12-41d2-983e-83e49d6df137\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.986642 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-config-data\") pod \"b8e3652a-3b12-41d2-983e-83e49d6df137\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.986674 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-logs\") pod \"b8e3652a-3b12-41d2-983e-83e49d6df137\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.986714 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"b8e3652a-3b12-41d2-983e-83e49d6df137\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.986752 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw2wn\" (UniqueName: \"kubernetes.io/projected/b8e3652a-3b12-41d2-983e-83e49d6df137-kube-api-access-bw2wn\") pod \"b8e3652a-3b12-41d2-983e-83e49d6df137\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.986851 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-httpd-run\") pod \"b8e3652a-3b12-41d2-983e-83e49d6df137\" (UID: \"b8e3652a-3b12-41d2-983e-83e49d6df137\") " Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.987661 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-logs" (OuterVolumeSpecName: "logs") pod "b8e3652a-3b12-41d2-983e-83e49d6df137" (UID: "b8e3652a-3b12-41d2-983e-83e49d6df137"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.987840 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b8e3652a-3b12-41d2-983e-83e49d6df137" (UID: "b8e3652a-3b12-41d2-983e-83e49d6df137"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:01 crc kubenswrapper[4715]: I1210 09:52:01.993804 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e3652a-3b12-41d2-983e-83e49d6df137-kube-api-access-bw2wn" (OuterVolumeSpecName: "kube-api-access-bw2wn") pod "b8e3652a-3b12-41d2-983e-83e49d6df137" (UID: "b8e3652a-3b12-41d2-983e-83e49d6df137"). InnerVolumeSpecName "kube-api-access-bw2wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:01.994491 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "b8e3652a-3b12-41d2-983e-83e49d6df137" (UID: "b8e3652a-3b12-41d2-983e-83e49d6df137"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:01.995069 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-scripts" (OuterVolumeSpecName: "scripts") pod "b8e3652a-3b12-41d2-983e-83e49d6df137" (UID: "b8e3652a-3b12-41d2-983e-83e49d6df137"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.021546 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8e3652a-3b12-41d2-983e-83e49d6df137" (UID: "b8e3652a-3b12-41d2-983e-83e49d6df137"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.037035 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b8e3652a-3b12-41d2-983e-83e49d6df137","Type":"ContainerDied","Data":"e676518c27c172350ddcbd19f95ff35ab0d2b23010375c050b24cff484e59e2b"} Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.037065 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.045701 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-config-data" (OuterVolumeSpecName: "config-data") pod "b8e3652a-3b12-41d2-983e-83e49d6df137" (UID: "b8e3652a-3b12-41d2-983e-83e49d6df137"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.088705 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.088735 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.088779 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.088790 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e3652a-3b12-41d2-983e-83e49d6df137-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.088801 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8e3652a-3b12-41d2-983e-83e49d6df137-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.088855 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.088867 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw2wn\" (UniqueName: \"kubernetes.io/projected/b8e3652a-3b12-41d2-983e-83e49d6df137-kube-api-access-bw2wn\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.110718 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.189939 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.387505 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.404635 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.430904 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:52:02 crc kubenswrapper[4715]: E1210 09:52:02.431451 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e3652a-3b12-41d2-983e-83e49d6df137" containerName="glance-httpd" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.431474 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e3652a-3b12-41d2-983e-83e49d6df137" containerName="glance-httpd" Dec 10 09:52:02 crc kubenswrapper[4715]: E1210 09:52:02.431495 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e3652a-3b12-41d2-983e-83e49d6df137" containerName="glance-log" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.431504 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e3652a-3b12-41d2-983e-83e49d6df137" containerName="glance-log" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.431764 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e3652a-3b12-41d2-983e-83e49d6df137" containerName="glance-log" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.431784 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e3652a-3b12-41d2-983e-83e49d6df137" containerName="glance-httpd" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.432963 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.437115 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.438007 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.447557 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.496893 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.496999 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.497032 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.497061 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzbp8\" (UniqueName: \"kubernetes.io/projected/207b4574-9a00-43c1-9d73-24e23343f35c-kube-api-access-vzbp8\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.497094 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.497123 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.497155 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-logs\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.497177 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.598574 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.598666 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.598696 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.598727 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzbp8\" (UniqueName: \"kubernetes.io/projected/207b4574-9a00-43c1-9d73-24e23343f35c-kube-api-access-vzbp8\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.598756 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.598785 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.598818 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-logs\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.598842 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.599062 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.599498 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.599556 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-logs\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.603552 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.603713 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.603764 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.607116 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.620297 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzbp8\" (UniqueName: \"kubernetes.io/projected/207b4574-9a00-43c1-9d73-24e23343f35c-kube-api-access-vzbp8\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.628636 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:52:02 crc kubenswrapper[4715]: I1210 09:52:02.753976 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:03 crc kubenswrapper[4715]: I1210 09:52:03.508231 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-d462h" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Dec 10 09:52:03 crc kubenswrapper[4715]: I1210 09:52:03.614454 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8e3652a-3b12-41d2-983e-83e49d6df137" path="/var/lib/kubelet/pods/b8e3652a-3b12-41d2-983e-83e49d6df137/volumes" Dec 10 09:52:06 crc kubenswrapper[4715]: E1210 09:52:06.489544 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 10 09:52:06 crc kubenswrapper[4715]: E1210 09:52:06.490031 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n564h94h67fhd8h84h66dh67h54bh584h8ch57ch55ch5c9h64bh5bfh664h6ch87h57bhcch69h64fh54chf6h588hcdhch598h557h67dhdbh67q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bhs26,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6948bcc975-bdnnf_openstack(be891085-7461-419d-82ae-f01a2643f6b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:52:06 crc kubenswrapper[4715]: E1210 09:52:06.492193 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 10 09:52:06 crc kubenswrapper[4715]: E1210 09:52:06.492270 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6948bcc975-bdnnf" podUID="be891085-7461-419d-82ae-f01a2643f6b4" Dec 10 09:52:06 crc kubenswrapper[4715]: E1210 09:52:06.492324 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n569h98hf8hf5h74h57fh668h645h655h56ch5bdhd4h85h5f4h54ch5c9h5f5h5d6h5dfh564h66bh79hddhfch67fh545h56hbch4hfch586h87q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s894m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7fcc456ff5-rc25b_openstack(5479ff20-2e76-4be0-807b-eca95c4cea9b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:52:06 crc kubenswrapper[4715]: E1210 09:52:06.494348 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7fcc456ff5-rc25b" podUID="5479ff20-2e76-4be0-807b-eca95c4cea9b" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.582724 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.691347 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-config-data\") pod \"465326ec-ce56-462a-9c90-ae7294e40b64\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.691475 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-credential-keys\") pod \"465326ec-ce56-462a-9c90-ae7294e40b64\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.691540 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65pt5\" (UniqueName: \"kubernetes.io/projected/465326ec-ce56-462a-9c90-ae7294e40b64-kube-api-access-65pt5\") pod \"465326ec-ce56-462a-9c90-ae7294e40b64\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.691600 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-combined-ca-bundle\") pod \"465326ec-ce56-462a-9c90-ae7294e40b64\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.691623 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-scripts\") pod \"465326ec-ce56-462a-9c90-ae7294e40b64\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.691794 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-fernet-keys\") pod \"465326ec-ce56-462a-9c90-ae7294e40b64\" (UID: \"465326ec-ce56-462a-9c90-ae7294e40b64\") " Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.698629 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-scripts" (OuterVolumeSpecName: "scripts") pod "465326ec-ce56-462a-9c90-ae7294e40b64" (UID: "465326ec-ce56-462a-9c90-ae7294e40b64"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.698657 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/465326ec-ce56-462a-9c90-ae7294e40b64-kube-api-access-65pt5" (OuterVolumeSpecName: "kube-api-access-65pt5") pod "465326ec-ce56-462a-9c90-ae7294e40b64" (UID: "465326ec-ce56-462a-9c90-ae7294e40b64"). InnerVolumeSpecName "kube-api-access-65pt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.698805 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "465326ec-ce56-462a-9c90-ae7294e40b64" (UID: "465326ec-ce56-462a-9c90-ae7294e40b64"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.700583 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "465326ec-ce56-462a-9c90-ae7294e40b64" (UID: "465326ec-ce56-462a-9c90-ae7294e40b64"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.720216 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "465326ec-ce56-462a-9c90-ae7294e40b64" (UID: "465326ec-ce56-462a-9c90-ae7294e40b64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.722394 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-config-data" (OuterVolumeSpecName: "config-data") pod "465326ec-ce56-462a-9c90-ae7294e40b64" (UID: "465326ec-ce56-462a-9c90-ae7294e40b64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.795290 4715 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.795328 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.795341 4715 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.795358 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65pt5\" (UniqueName: \"kubernetes.io/projected/465326ec-ce56-462a-9c90-ae7294e40b64-kube-api-access-65pt5\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.795368 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:06 crc kubenswrapper[4715]: I1210 09:52:06.795375 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/465326ec-ce56-462a-9c90-ae7294e40b64-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.095144 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7hv68" event={"ID":"465326ec-ce56-462a-9c90-ae7294e40b64","Type":"ContainerDied","Data":"2ef50d81e4152f179b7219e03c228c0d308f2fb3a9ffba5c0c51d27b2e2b4996"} Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.095198 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ef50d81e4152f179b7219e03c228c0d308f2fb3a9ffba5c0c51d27b2e2b4996" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.095257 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7hv68" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.676739 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-7hv68"] Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.689032 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-7hv68"] Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.774931 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-x2v2z"] Dec 10 09:52:07 crc kubenswrapper[4715]: E1210 09:52:07.775441 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="465326ec-ce56-462a-9c90-ae7294e40b64" containerName="keystone-bootstrap" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.775465 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="465326ec-ce56-462a-9c90-ae7294e40b64" containerName="keystone-bootstrap" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.775703 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="465326ec-ce56-462a-9c90-ae7294e40b64" containerName="keystone-bootstrap" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.776416 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.779075 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.779480 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.779705 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.780081 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.784479 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-x2v2z"] Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.796879 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-csxkd" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.827814 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-config-data\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.828009 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjrn5\" (UniqueName: \"kubernetes.io/projected/27133872-3a45-45a6-9d91-f645c0c9b264-kube-api-access-bjrn5\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.828288 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-combined-ca-bundle\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.828389 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-credential-keys\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.828421 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-fernet-keys\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.828476 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-scripts\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.930122 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjrn5\" (UniqueName: \"kubernetes.io/projected/27133872-3a45-45a6-9d91-f645c0c9b264-kube-api-access-bjrn5\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.930533 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-combined-ca-bundle\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.930888 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-credential-keys\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.931476 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-fernet-keys\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.931573 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-scripts\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.931610 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-config-data\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.938713 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-scripts\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.938979 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-combined-ca-bundle\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.939614 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-credential-keys\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.940009 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-config-data\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.946669 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-fernet-keys\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:07 crc kubenswrapper[4715]: I1210 09:52:07.950556 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjrn5\" (UniqueName: \"kubernetes.io/projected/27133872-3a45-45a6-9d91-f645c0c9b264-kube-api-access-bjrn5\") pod \"keystone-bootstrap-x2v2z\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:08 crc kubenswrapper[4715]: I1210 09:52:08.110526 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:09 crc kubenswrapper[4715]: I1210 09:52:09.113084 4715 generic.go:334] "Generic (PLEG): container finished" podID="a08f4204-23b4-425a-a116-cfd37741c87f" containerID="0c4f3f531995496685f907c8d2b13f8583be7b428dd7c227e2622600162bb714" exitCode=0 Dec 10 09:52:09 crc kubenswrapper[4715]: I1210 09:52:09.113137 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2vj6t" event={"ID":"a08f4204-23b4-425a-a116-cfd37741c87f","Type":"ContainerDied","Data":"0c4f3f531995496685f907c8d2b13f8583be7b428dd7c227e2622600162bb714"} Dec 10 09:52:09 crc kubenswrapper[4715]: I1210 09:52:09.620079 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="465326ec-ce56-462a-9c90-ae7294e40b64" path="/var/lib/kubelet/pods/465326ec-ce56-462a-9c90-ae7294e40b64/volumes" Dec 10 09:52:13 crc kubenswrapper[4715]: I1210 09:52:13.508155 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-d462h" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Dec 10 09:52:13 crc kubenswrapper[4715]: I1210 09:52:13.508899 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:52:15 crc kubenswrapper[4715]: E1210 09:52:15.548232 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 10 09:52:15 crc kubenswrapper[4715]: E1210 09:52:15.548705 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chfh65bh5cchf8h697hdh54dh56h7chf5h675h5bbh594h9fh67dhfdh65ch587h54ch5bbh77h85h5f8h4hc8h66h588h96h56ch5d6h664q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p2jmb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-59f89569c5-knqlx_openstack(90e59bef-1c16-4d24-800d-8bac5cb3e990): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:52:15 crc kubenswrapper[4715]: E1210 09:52:15.553439 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-59f89569c5-knqlx" podUID="90e59bef-1c16-4d24-800d-8bac5cb3e990" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.236216 4715 scope.go:117] "RemoveContainer" containerID="91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.343783 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.354162 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.364903 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.374173 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.433652 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-config-data\") pod \"be891085-7461-419d-82ae-f01a2643f6b4\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.433701 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-config\") pod \"a08f4204-23b4-425a-a116-cfd37741c87f\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.433734 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-combined-ca-bundle\") pod \"a08f4204-23b4-425a-a116-cfd37741c87f\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.433762 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-config-data\") pod \"5479ff20-2e76-4be0-807b-eca95c4cea9b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.433793 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-scripts\") pod \"be891085-7461-419d-82ae-f01a2643f6b4\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.433839 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kn6lm\" (UniqueName: \"kubernetes.io/projected/a08f4204-23b4-425a-a116-cfd37741c87f-kube-api-access-kn6lm\") pod \"a08f4204-23b4-425a-a116-cfd37741c87f\" (UID: \"a08f4204-23b4-425a-a116-cfd37741c87f\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.433888 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be891085-7461-419d-82ae-f01a2643f6b4-logs\") pod \"be891085-7461-419d-82ae-f01a2643f6b4\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.434019 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5479ff20-2e76-4be0-807b-eca95c4cea9b-logs\") pod \"5479ff20-2e76-4be0-807b-eca95c4cea9b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.434049 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/be891085-7461-419d-82ae-f01a2643f6b4-horizon-secret-key\") pod \"be891085-7461-419d-82ae-f01a2643f6b4\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.434087 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s894m\" (UniqueName: \"kubernetes.io/projected/5479ff20-2e76-4be0-807b-eca95c4cea9b-kube-api-access-s894m\") pod \"5479ff20-2e76-4be0-807b-eca95c4cea9b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.434104 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhs26\" (UniqueName: \"kubernetes.io/projected/be891085-7461-419d-82ae-f01a2643f6b4-kube-api-access-bhs26\") pod \"be891085-7461-419d-82ae-f01a2643f6b4\" (UID: \"be891085-7461-419d-82ae-f01a2643f6b4\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.434125 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5479ff20-2e76-4be0-807b-eca95c4cea9b-horizon-secret-key\") pod \"5479ff20-2e76-4be0-807b-eca95c4cea9b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.434205 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-scripts\") pod \"5479ff20-2e76-4be0-807b-eca95c4cea9b\" (UID: \"5479ff20-2e76-4be0-807b-eca95c4cea9b\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.434626 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be891085-7461-419d-82ae-f01a2643f6b4-logs" (OuterVolumeSpecName: "logs") pod "be891085-7461-419d-82ae-f01a2643f6b4" (UID: "be891085-7461-419d-82ae-f01a2643f6b4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.435102 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-scripts" (OuterVolumeSpecName: "scripts") pod "be891085-7461-419d-82ae-f01a2643f6b4" (UID: "be891085-7461-419d-82ae-f01a2643f6b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.435265 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-scripts" (OuterVolumeSpecName: "scripts") pod "5479ff20-2e76-4be0-807b-eca95c4cea9b" (UID: "5479ff20-2e76-4be0-807b-eca95c4cea9b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.435881 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-config-data" (OuterVolumeSpecName: "config-data") pod "5479ff20-2e76-4be0-807b-eca95c4cea9b" (UID: "5479ff20-2e76-4be0-807b-eca95c4cea9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.436604 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5479ff20-2e76-4be0-807b-eca95c4cea9b-logs" (OuterVolumeSpecName: "logs") pod "5479ff20-2e76-4be0-807b-eca95c4cea9b" (UID: "5479ff20-2e76-4be0-807b-eca95c4cea9b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.438408 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-config-data" (OuterVolumeSpecName: "config-data") pod "be891085-7461-419d-82ae-f01a2643f6b4" (UID: "be891085-7461-419d-82ae-f01a2643f6b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.447198 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5479ff20-2e76-4be0-807b-eca95c4cea9b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5479ff20-2e76-4be0-807b-eca95c4cea9b" (UID: "5479ff20-2e76-4be0-807b-eca95c4cea9b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.447237 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5479ff20-2e76-4be0-807b-eca95c4cea9b-kube-api-access-s894m" (OuterVolumeSpecName: "kube-api-access-s894m") pod "5479ff20-2e76-4be0-807b-eca95c4cea9b" (UID: "5479ff20-2e76-4be0-807b-eca95c4cea9b"). InnerVolumeSpecName "kube-api-access-s894m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.447335 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be891085-7461-419d-82ae-f01a2643f6b4-kube-api-access-bhs26" (OuterVolumeSpecName: "kube-api-access-bhs26") pod "be891085-7461-419d-82ae-f01a2643f6b4" (UID: "be891085-7461-419d-82ae-f01a2643f6b4"). InnerVolumeSpecName "kube-api-access-bhs26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.447409 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a08f4204-23b4-425a-a116-cfd37741c87f-kube-api-access-kn6lm" (OuterVolumeSpecName: "kube-api-access-kn6lm") pod "a08f4204-23b4-425a-a116-cfd37741c87f" (UID: "a08f4204-23b4-425a-a116-cfd37741c87f"). InnerVolumeSpecName "kube-api-access-kn6lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.454199 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be891085-7461-419d-82ae-f01a2643f6b4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "be891085-7461-419d-82ae-f01a2643f6b4" (UID: "be891085-7461-419d-82ae-f01a2643f6b4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.465107 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-config" (OuterVolumeSpecName: "config") pod "a08f4204-23b4-425a-a116-cfd37741c87f" (UID: "a08f4204-23b4-425a-a116-cfd37741c87f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.480484 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a08f4204-23b4-425a-a116-cfd37741c87f" (UID: "a08f4204-23b4-425a-a116-cfd37741c87f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.536321 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-sb\") pod \"de02f9b6-491d-4be4-9f06-7c1c53976694\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.536534 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwz4r\" (UniqueName: \"kubernetes.io/projected/de02f9b6-491d-4be4-9f06-7c1c53976694-kube-api-access-rwz4r\") pod \"de02f9b6-491d-4be4-9f06-7c1c53976694\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.536590 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-config\") pod \"de02f9b6-491d-4be4-9f06-7c1c53976694\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.536638 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-dns-svc\") pod \"de02f9b6-491d-4be4-9f06-7c1c53976694\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.536684 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-nb\") pod \"de02f9b6-491d-4be4-9f06-7c1c53976694\" (UID: \"de02f9b6-491d-4be4-9f06-7c1c53976694\") " Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537206 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5479ff20-2e76-4be0-807b-eca95c4cea9b-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537231 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/be891085-7461-419d-82ae-f01a2643f6b4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537247 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s894m\" (UniqueName: \"kubernetes.io/projected/5479ff20-2e76-4be0-807b-eca95c4cea9b-kube-api-access-s894m\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537259 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhs26\" (UniqueName: \"kubernetes.io/projected/be891085-7461-419d-82ae-f01a2643f6b4-kube-api-access-bhs26\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537271 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5479ff20-2e76-4be0-807b-eca95c4cea9b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537311 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537323 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537335 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537346 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a08f4204-23b4-425a-a116-cfd37741c87f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537356 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5479ff20-2e76-4be0-807b-eca95c4cea9b-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537367 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/be891085-7461-419d-82ae-f01a2643f6b4-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537378 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kn6lm\" (UniqueName: \"kubernetes.io/projected/a08f4204-23b4-425a-a116-cfd37741c87f-kube-api-access-kn6lm\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.537389 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be891085-7461-419d-82ae-f01a2643f6b4-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.541874 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de02f9b6-491d-4be4-9f06-7c1c53976694-kube-api-access-rwz4r" (OuterVolumeSpecName: "kube-api-access-rwz4r") pod "de02f9b6-491d-4be4-9f06-7c1c53976694" (UID: "de02f9b6-491d-4be4-9f06-7c1c53976694"). InnerVolumeSpecName "kube-api-access-rwz4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.580589 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-config" (OuterVolumeSpecName: "config") pod "de02f9b6-491d-4be4-9f06-7c1c53976694" (UID: "de02f9b6-491d-4be4-9f06-7c1c53976694"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.584729 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de02f9b6-491d-4be4-9f06-7c1c53976694" (UID: "de02f9b6-491d-4be4-9f06-7c1c53976694"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.585143 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de02f9b6-491d-4be4-9f06-7c1c53976694" (UID: "de02f9b6-491d-4be4-9f06-7c1c53976694"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.586395 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de02f9b6-491d-4be4-9f06-7c1c53976694" (UID: "de02f9b6-491d-4be4-9f06-7c1c53976694"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.638806 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwz4r\" (UniqueName: \"kubernetes.io/projected/de02f9b6-491d-4be4-9f06-7c1c53976694-kube-api-access-rwz4r\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.638849 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.638864 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.638876 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:17 crc kubenswrapper[4715]: I1210 09:52:17.638887 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de02f9b6-491d-4be4-9f06-7c1c53976694-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.199216 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2vj6t" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.199209 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2vj6t" event={"ID":"a08f4204-23b4-425a-a116-cfd37741c87f","Type":"ContainerDied","Data":"ea9785ad784790c9925ffbeee6c2aa38af92552cd0171c9f66388b1cac862126"} Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.199382 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea9785ad784790c9925ffbeee6c2aa38af92552cd0171c9f66388b1cac862126" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.200415 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6948bcc975-bdnnf" event={"ID":"be891085-7461-419d-82ae-f01a2643f6b4","Type":"ContainerDied","Data":"14915d715ef83133cb338df20bed8f993cf984ebded8b0ce6deae033f0944536"} Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.200427 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6948bcc975-bdnnf" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.201452 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fcc456ff5-rc25b" event={"ID":"5479ff20-2e76-4be0-807b-eca95c4cea9b","Type":"ContainerDied","Data":"a2a1703f47703be70719d2cb495de52284213e1863097341da7a601973029779"} Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.201508 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fcc456ff5-rc25b" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.205908 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-d462h" event={"ID":"de02f9b6-491d-4be4-9f06-7c1c53976694","Type":"ContainerDied","Data":"fb8d87cfcd242f8575387dd7b6247e9f00d21c72d89c7fdb5ec1629c08d252d3"} Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.205959 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-d462h" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.263223 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fcc456ff5-rc25b"] Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.271107 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7fcc456ff5-rc25b"] Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.327176 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6948bcc975-bdnnf"] Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.334432 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6948bcc975-bdnnf"] Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.344669 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-d462h"] Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.360684 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-d462h"] Dec 10 09:52:18 crc kubenswrapper[4715]: E1210 09:52:18.386750 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 10 09:52:18 crc kubenswrapper[4715]: E1210 09:52:18.386932 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dwm8t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-28hsp_openstack(8bf72f3c-8033-4018-b330-72fa2774f402): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:52:18 crc kubenswrapper[4715]: E1210 09:52:18.388068 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-28hsp" podUID="8bf72f3c-8033-4018-b330-72fa2774f402" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.453861 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.509139 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-d462h" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.560153 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-config-data\") pod \"90e59bef-1c16-4d24-800d-8bac5cb3e990\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.560209 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2jmb\" (UniqueName: \"kubernetes.io/projected/90e59bef-1c16-4d24-800d-8bac5cb3e990-kube-api-access-p2jmb\") pod \"90e59bef-1c16-4d24-800d-8bac5cb3e990\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.560309 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-scripts\") pod \"90e59bef-1c16-4d24-800d-8bac5cb3e990\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.560338 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90e59bef-1c16-4d24-800d-8bac5cb3e990-logs\") pod \"90e59bef-1c16-4d24-800d-8bac5cb3e990\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.560495 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90e59bef-1c16-4d24-800d-8bac5cb3e990-horizon-secret-key\") pod \"90e59bef-1c16-4d24-800d-8bac5cb3e990\" (UID: \"90e59bef-1c16-4d24-800d-8bac5cb3e990\") " Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.560813 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-config-data" (OuterVolumeSpecName: "config-data") pod "90e59bef-1c16-4d24-800d-8bac5cb3e990" (UID: "90e59bef-1c16-4d24-800d-8bac5cb3e990"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.561055 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.561328 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-scripts" (OuterVolumeSpecName: "scripts") pod "90e59bef-1c16-4d24-800d-8bac5cb3e990" (UID: "90e59bef-1c16-4d24-800d-8bac5cb3e990"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.562168 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90e59bef-1c16-4d24-800d-8bac5cb3e990-logs" (OuterVolumeSpecName: "logs") pod "90e59bef-1c16-4d24-800d-8bac5cb3e990" (UID: "90e59bef-1c16-4d24-800d-8bac5cb3e990"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.565893 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90e59bef-1c16-4d24-800d-8bac5cb3e990-kube-api-access-p2jmb" (OuterVolumeSpecName: "kube-api-access-p2jmb") pod "90e59bef-1c16-4d24-800d-8bac5cb3e990" (UID: "90e59bef-1c16-4d24-800d-8bac5cb3e990"). InnerVolumeSpecName "kube-api-access-p2jmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.567873 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90e59bef-1c16-4d24-800d-8bac5cb3e990-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "90e59bef-1c16-4d24-800d-8bac5cb3e990" (UID: "90e59bef-1c16-4d24-800d-8bac5cb3e990"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.574187 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-t8855"] Dec 10 09:52:18 crc kubenswrapper[4715]: E1210 09:52:18.574549 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a08f4204-23b4-425a-a116-cfd37741c87f" containerName="neutron-db-sync" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.574565 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a08f4204-23b4-425a-a116-cfd37741c87f" containerName="neutron-db-sync" Dec 10 09:52:18 crc kubenswrapper[4715]: E1210 09:52:18.574584 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerName="dnsmasq-dns" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.574590 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerName="dnsmasq-dns" Dec 10 09:52:18 crc kubenswrapper[4715]: E1210 09:52:18.574605 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerName="init" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.574611 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerName="init" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.574859 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a08f4204-23b4-425a-a116-cfd37741c87f" containerName="neutron-db-sync" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.574871 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" containerName="dnsmasq-dns" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.580700 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.593243 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-t8855"] Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.663139 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.664595 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-config\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.664641 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.664673 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.664710 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.664897 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb76t\" (UniqueName: \"kubernetes.io/projected/296b014c-dc51-422d-ab1a-3e2018759e4a-kube-api-access-jb76t\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.665077 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2jmb\" (UniqueName: \"kubernetes.io/projected/90e59bef-1c16-4d24-800d-8bac5cb3e990-kube-api-access-p2jmb\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.665090 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90e59bef-1c16-4d24-800d-8bac5cb3e990-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.665099 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90e59bef-1c16-4d24-800d-8bac5cb3e990-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.665110 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/90e59bef-1c16-4d24-800d-8bac5cb3e990-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.680511 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6c77448d4b-7p5j8"] Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.691598 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c77448d4b-7p5j8"] Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.691760 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.693374 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.693794 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.694709 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.695137 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qnst4" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766387 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-ovndb-tls-certs\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766497 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-httpd-config\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766546 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-config\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766579 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766607 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766628 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766707 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-config\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766741 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfjvb\" (UniqueName: \"kubernetes.io/projected/96bdc06f-533b-4ba3-8074-a92466586d44-kube-api-access-vfjvb\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766783 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb76t\" (UniqueName: \"kubernetes.io/projected/296b014c-dc51-422d-ab1a-3e2018759e4a-kube-api-access-jb76t\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.766827 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-combined-ca-bundle\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.768652 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.768663 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.768990 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.769418 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-config\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.769554 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.811707 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb76t\" (UniqueName: \"kubernetes.io/projected/296b014c-dc51-422d-ab1a-3e2018759e4a-kube-api-access-jb76t\") pod \"dnsmasq-dns-55f844cf75-t8855\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.870039 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-config\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.870106 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfjvb\" (UniqueName: \"kubernetes.io/projected/96bdc06f-533b-4ba3-8074-a92466586d44-kube-api-access-vfjvb\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.870176 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-combined-ca-bundle\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.870226 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-ovndb-tls-certs\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.870285 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-httpd-config\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.875818 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-combined-ca-bundle\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.879624 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-httpd-config\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.880403 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-ovndb-tls-certs\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.880544 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-config\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.896688 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfjvb\" (UniqueName: \"kubernetes.io/projected/96bdc06f-533b-4ba3-8074-a92466586d44-kube-api-access-vfjvb\") pod \"neutron-6c77448d4b-7p5j8\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:18 crc kubenswrapper[4715]: I1210 09:52:18.962616 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.010851 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.220342 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59f89569c5-knqlx" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.221112 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59f89569c5-knqlx" event={"ID":"90e59bef-1c16-4d24-800d-8bac5cb3e990","Type":"ContainerDied","Data":"c161b0e868272e83c4ee1c3ca7ed7f7829257b735b3c5861020bb79d41b71a6a"} Dec 10 09:52:19 crc kubenswrapper[4715]: E1210 09:52:19.223551 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-28hsp" podUID="8bf72f3c-8033-4018-b330-72fa2774f402" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.287982 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59f89569c5-knqlx"] Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.295086 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-59f89569c5-knqlx"] Dec 10 09:52:19 crc kubenswrapper[4715]: E1210 09:52:19.514268 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 10 09:52:19 crc kubenswrapper[4715]: E1210 09:52:19.514424 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hz8v7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-gzr48_openstack(66856339-f4ab-4102-93a8-adc952877ccd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 09:52:19 crc kubenswrapper[4715]: E1210 09:52:19.515621 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-gzr48" podUID="66856339-f4ab-4102-93a8-adc952877ccd" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.626144 4715 scope.go:117] "RemoveContainer" containerID="0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043" Dec 10 09:52:19 crc kubenswrapper[4715]: E1210 09:52:19.627019 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043\": container with ID starting with 0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043 not found: ID does not exist" containerID="0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.627050 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043"} err="failed to get container status \"0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043\": rpc error: code = NotFound desc = could not find container \"0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043\": container with ID starting with 0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043 not found: ID does not exist" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.627075 4715 scope.go:117] "RemoveContainer" containerID="91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461" Dec 10 09:52:19 crc kubenswrapper[4715]: E1210 09:52:19.628025 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461\": container with ID starting with 91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461 not found: ID does not exist" containerID="91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.628235 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461"} err="failed to get container status \"91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461\": rpc error: code = NotFound desc = could not find container \"91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461\": container with ID starting with 91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461 not found: ID does not exist" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.628263 4715 scope.go:117] "RemoveContainer" containerID="0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.631945 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043"} err="failed to get container status \"0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043\": rpc error: code = NotFound desc = could not find container \"0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043\": container with ID starting with 0e045091d356088391f1a76d8324cfd40678f74be2a242da49688e656a505043 not found: ID does not exist" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.632041 4715 scope.go:117] "RemoveContainer" containerID="91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.633257 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461"} err="failed to get container status \"91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461\": rpc error: code = NotFound desc = could not find container \"91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461\": container with ID starting with 91de7916c2d340fdd15a26bbd2dc436d87590166bd94c64e06a08e310ef46461 not found: ID does not exist" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.633281 4715 scope.go:117] "RemoveContainer" containerID="c5aeac5a1f59add4398e69a067c7f3206bbe2c372ba9ca2753e784a5728eeb51" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.641875 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5479ff20-2e76-4be0-807b-eca95c4cea9b" path="/var/lib/kubelet/pods/5479ff20-2e76-4be0-807b-eca95c4cea9b/volumes" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.642473 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90e59bef-1c16-4d24-800d-8bac5cb3e990" path="/var/lib/kubelet/pods/90e59bef-1c16-4d24-800d-8bac5cb3e990/volumes" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.643600 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be891085-7461-419d-82ae-f01a2643f6b4" path="/var/lib/kubelet/pods/be891085-7461-419d-82ae-f01a2643f6b4/volumes" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.644727 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de02f9b6-491d-4be4-9f06-7c1c53976694" path="/var/lib/kubelet/pods/de02f9b6-491d-4be4-9f06-7c1c53976694/volumes" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.822084 4715 scope.go:117] "RemoveContainer" containerID="63db12bc74a53c2de57ce51c452d6033c15a94bc92a94f9eb8aa4e6a25903b05" Dec 10 09:52:19 crc kubenswrapper[4715]: I1210 09:52:19.988507 4715 scope.go:117] "RemoveContainer" containerID="82c73dd9e350a16ff73aac255bc07d2d8de274b87dacef3429c9d880d308c897" Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.039973 4715 scope.go:117] "RemoveContainer" containerID="bce68d6d483eb4de7af24eedf6792720982d1fdd5aa1661295c9e6ac5f66eaf9" Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.126305 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.237251 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-857b4dfcd4-wnlbz"] Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.260277 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67f6cc7984-tshct"] Dec 10 09:52:20 crc kubenswrapper[4715]: W1210 09:52:20.285283 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4214260f_423f_47fb_b062_c752f3519175.slice/crio-1e448d82cb6a4d6ceb85ee8b5e419b516d4934260285a399b58c7775d3c412a1 WatchSource:0}: Error finding container 1e448d82cb6a4d6ceb85ee8b5e419b516d4934260285a399b58c7775d3c412a1: Status 404 returned error can't find the container with id 1e448d82cb6a4d6ceb85ee8b5e419b516d4934260285a399b58c7775d3c412a1 Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.334404 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zq2xt" event={"ID":"53507dfd-6f17-486f-8770-c073ec1b42a3","Type":"ContainerStarted","Data":"c51c3b816413cb9c1bd195a8e5ae598869eacdc3310e4b370b5d49c164de15ad"} Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.365032 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2167d98b-0365-421e-9d3d-6ff62f7bfb61","Type":"ContainerStarted","Data":"382c1037c3c28ee70ddc99fb57d441e603a2485e73eef3c7410495d5fb30133b"} Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.365874 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7c0209db-d227-4d43-a132-e2393ce0d5c3","Type":"ContainerStarted","Data":"9f4f99ad8eb11aa447f1bce8ab9c0bbcc65324a07a2a8c0afec035773111102d"} Dec 10 09:52:20 crc kubenswrapper[4715]: E1210 09:52:20.368749 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-gzr48" podUID="66856339-f4ab-4102-93a8-adc952877ccd" Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.374156 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-zq2xt" podStartSLOduration=5.767931965 podStartE2EDuration="35.374132659s" podCreationTimestamp="2025-12-10 09:51:45 +0000 UTC" firstStartedPulling="2025-12-10 09:51:47.635526486 +0000 UTC m=+1070.379072737" lastFinishedPulling="2025-12-10 09:52:17.24172718 +0000 UTC m=+1099.985273431" observedRunningTime="2025-12-10 09:52:20.358481056 +0000 UTC m=+1103.102027307" watchObservedRunningTime="2025-12-10 09:52:20.374132659 +0000 UTC m=+1103.117678910" Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.427050 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-t8855"] Dec 10 09:52:20 crc kubenswrapper[4715]: W1210 09:52:20.433615 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod296b014c_dc51_422d_ab1a_3e2018759e4a.slice/crio-84db2e84a14905e893ae687f3ce9ced9a715c8ef9230e64405add47dd1075e80 WatchSource:0}: Error finding container 84db2e84a14905e893ae687f3ce9ced9a715c8ef9230e64405add47dd1075e80: Status 404 returned error can't find the container with id 84db2e84a14905e893ae687f3ce9ced9a715c8ef9230e64405add47dd1075e80 Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.434899 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-x2v2z"] Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.506678 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:52:20 crc kubenswrapper[4715]: I1210 09:52:20.577884 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c77448d4b-7p5j8"] Dec 10 09:52:20 crc kubenswrapper[4715]: W1210 09:52:20.592562 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96bdc06f_533b_4ba3_8074_a92466586d44.slice/crio-7dee8d82e7b9b5c6a3321400f48ca87b56e22c283c551b61c4a9fdbc01bc8e4f WatchSource:0}: Error finding container 7dee8d82e7b9b5c6a3321400f48ca87b56e22c283c551b61c4a9fdbc01bc8e4f: Status 404 returned error can't find the container with id 7dee8d82e7b9b5c6a3321400f48ca87b56e22c283c551b61c4a9fdbc01bc8e4f Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.306903 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-85df69bf85-6cwpq"] Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.309061 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.311699 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.311848 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.318321 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85df69bf85-6cwpq"] Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.363702 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-internal-tls-certs\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.363796 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-combined-ca-bundle\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.363832 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2cjb\" (UniqueName: \"kubernetes.io/projected/0cf58c4f-3594-4b20-a3fb-313c0f25a844-kube-api-access-d2cjb\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.363855 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-config\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.363889 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-ovndb-tls-certs\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.363924 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-httpd-config\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.363960 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-public-tls-certs\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.394716 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-857b4dfcd4-wnlbz" event={"ID":"4214260f-423f-47fb-b062-c752f3519175","Type":"ContainerStarted","Data":"d4d2de91913dd2ebf75948dc14b51a202962a866a936f9ce143d9542a522787e"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.394782 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-857b4dfcd4-wnlbz" event={"ID":"4214260f-423f-47fb-b062-c752f3519175","Type":"ContainerStarted","Data":"1e448d82cb6a4d6ceb85ee8b5e419b516d4934260285a399b58c7775d3c412a1"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.396492 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"207b4574-9a00-43c1-9d73-24e23343f35c","Type":"ContainerStarted","Data":"e3e5c7590b66d8da8450aeb403df0e3793c4bce42848ff430c0ae650fe038bf5"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.402824 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c77448d4b-7p5j8" event={"ID":"96bdc06f-533b-4ba3-8074-a92466586d44","Type":"ContainerStarted","Data":"f3063c5b676bdfc3fdb327164846de92265a2df2487ea64ff436fc663bdc6360"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.402867 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c77448d4b-7p5j8" event={"ID":"96bdc06f-533b-4ba3-8074-a92466586d44","Type":"ContainerStarted","Data":"06577babfec20c9050142133b19e2ff06bc98a9bdd42bed1ac09c8b05f0f3ed8"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.402878 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c77448d4b-7p5j8" event={"ID":"96bdc06f-533b-4ba3-8074-a92466586d44","Type":"ContainerStarted","Data":"7dee8d82e7b9b5c6a3321400f48ca87b56e22c283c551b61c4a9fdbc01bc8e4f"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.403341 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.410127 4715 generic.go:334] "Generic (PLEG): container finished" podID="296b014c-dc51-422d-ab1a-3e2018759e4a" containerID="34e5c1b0be35fb91bfd6fc9c791e3e819c821cd7fd272cab546dde7efdec1e10" exitCode=0 Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.410230 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-t8855" event={"ID":"296b014c-dc51-422d-ab1a-3e2018759e4a","Type":"ContainerDied","Data":"34e5c1b0be35fb91bfd6fc9c791e3e819c821cd7fd272cab546dde7efdec1e10"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.410260 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-t8855" event={"ID":"296b014c-dc51-422d-ab1a-3e2018759e4a","Type":"ContainerStarted","Data":"84db2e84a14905e893ae687f3ce9ced9a715c8ef9230e64405add47dd1075e80"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.414074 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f6cc7984-tshct" event={"ID":"5a1c5abd-cfc5-427f-a378-cef6d31deb0e","Type":"ContainerStarted","Data":"c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.414113 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f6cc7984-tshct" event={"ID":"5a1c5abd-cfc5-427f-a378-cef6d31deb0e","Type":"ContainerStarted","Data":"1ece62e33ef77421cffb2e5b7d6c6ffa3f2dbacaca2f6365b9fbc5b4bb8fe542"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.416167 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x2v2z" event={"ID":"27133872-3a45-45a6-9d91-f645c0c9b264","Type":"ContainerStarted","Data":"9ef9793a46d67594228c5ad70a237a786149e60b2da549a0aad6cb2b5bcc7930"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.416189 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x2v2z" event={"ID":"27133872-3a45-45a6-9d91-f645c0c9b264","Type":"ContainerStarted","Data":"c38a582938016190564b921cc60a17793d2e4debe2d26c5cf1dfe6afbbbcc83c"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.422213 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7c0209db-d227-4d43-a132-e2393ce0d5c3","Type":"ContainerStarted","Data":"a6a3751858a858b59e78c6e794614c1bb47d642d5cd6c2c664dee03ecd418eb9"} Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.441877 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6c77448d4b-7p5j8" podStartSLOduration=3.441855813 podStartE2EDuration="3.441855813s" podCreationTimestamp="2025-12-10 09:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:21.426667743 +0000 UTC m=+1104.170214004" watchObservedRunningTime="2025-12-10 09:52:21.441855813 +0000 UTC m=+1104.185402064" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.465163 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-public-tls-certs\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.465680 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-internal-tls-certs\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.465769 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-combined-ca-bundle\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.465810 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2cjb\" (UniqueName: \"kubernetes.io/projected/0cf58c4f-3594-4b20-a3fb-313c0f25a844-kube-api-access-d2cjb\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.465836 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-config\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.465880 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-ovndb-tls-certs\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.465903 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-httpd-config\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.471558 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-public-tls-certs\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.476048 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-x2v2z" podStartSLOduration=14.476031461 podStartE2EDuration="14.476031461s" podCreationTimestamp="2025-12-10 09:52:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:21.466381888 +0000 UTC m=+1104.209928139" watchObservedRunningTime="2025-12-10 09:52:21.476031461 +0000 UTC m=+1104.219577702" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.478373 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-config\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.483802 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-ovndb-tls-certs\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.491390 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-combined-ca-bundle\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.494719 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-httpd-config\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.511973 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cf58c4f-3594-4b20-a3fb-313c0f25a844-internal-tls-certs\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.522574 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2cjb\" (UniqueName: \"kubernetes.io/projected/0cf58c4f-3594-4b20-a3fb-313c0f25a844-kube-api-access-d2cjb\") pod \"neutron-85df69bf85-6cwpq\" (UID: \"0cf58c4f-3594-4b20-a3fb-313c0f25a844\") " pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:21 crc kubenswrapper[4715]: I1210 09:52:21.670468 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.463031 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2167d98b-0365-421e-9d3d-6ff62f7bfb61","Type":"ContainerStarted","Data":"d6a23411f595071e43d6ed0df1b67d55411b62de11ff04597a980928622d1e76"} Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.468675 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85df69bf85-6cwpq"] Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.473585 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7c0209db-d227-4d43-a132-e2393ce0d5c3","Type":"ContainerStarted","Data":"c3a435ba5c7d59aec6834021bd7ea496b620189b0bcf79a4db85af3b4e6c81de"} Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.473775 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7c0209db-d227-4d43-a132-e2393ce0d5c3" containerName="glance-log" containerID="cri-o://a6a3751858a858b59e78c6e794614c1bb47d642d5cd6c2c664dee03ecd418eb9" gracePeriod=30 Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.474532 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7c0209db-d227-4d43-a132-e2393ce0d5c3" containerName="glance-httpd" containerID="cri-o://c3a435ba5c7d59aec6834021bd7ea496b620189b0bcf79a4db85af3b4e6c81de" gracePeriod=30 Dec 10 09:52:22 crc kubenswrapper[4715]: W1210 09:52:22.489166 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cf58c4f_3594_4b20_a3fb_313c0f25a844.slice/crio-9d99b4602dca737ce57581d909b2632cacece949c76bac94ae734df97c6d1d22 WatchSource:0}: Error finding container 9d99b4602dca737ce57581d909b2632cacece949c76bac94ae734df97c6d1d22: Status 404 returned error can't find the container with id 9d99b4602dca737ce57581d909b2632cacece949c76bac94ae734df97c6d1d22 Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.490776 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-857b4dfcd4-wnlbz" event={"ID":"4214260f-423f-47fb-b062-c752f3519175","Type":"ContainerStarted","Data":"187441f4712046c726b7103f564385263b9a44def6e43a3b0e7d20c86c405704"} Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.499584 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=29.499570074 podStartE2EDuration="29.499570074s" podCreationTimestamp="2025-12-10 09:51:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:22.496689453 +0000 UTC m=+1105.240235704" watchObservedRunningTime="2025-12-10 09:52:22.499570074 +0000 UTC m=+1105.243116315" Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.499751 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"207b4574-9a00-43c1-9d73-24e23343f35c","Type":"ContainerStarted","Data":"d3d0255b6696448d82f647922157d65706d9eeb197b2dab337d7f96103ad86b1"} Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.508311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-t8855" event={"ID":"296b014c-dc51-422d-ab1a-3e2018759e4a","Type":"ContainerStarted","Data":"77d73f827aff7bd2ee4d514c4e9756d7255049a02f1d032a89c5b6b1b046e946"} Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.508450 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.513773 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f6cc7984-tshct" event={"ID":"5a1c5abd-cfc5-427f-a378-cef6d31deb0e","Type":"ContainerStarted","Data":"e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d"} Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.559592 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-857b4dfcd4-wnlbz" podStartSLOduration=26.08785884 podStartE2EDuration="26.559572463s" podCreationTimestamp="2025-12-10 09:51:56 +0000 UTC" firstStartedPulling="2025-12-10 09:52:20.298836358 +0000 UTC m=+1103.042382609" lastFinishedPulling="2025-12-10 09:52:20.770549981 +0000 UTC m=+1103.514096232" observedRunningTime="2025-12-10 09:52:22.545966168 +0000 UTC m=+1105.289512429" watchObservedRunningTime="2025-12-10 09:52:22.559572463 +0000 UTC m=+1105.303118714" Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.576140 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-t8855" podStartSLOduration=4.576118071 podStartE2EDuration="4.576118071s" podCreationTimestamp="2025-12-10 09:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:22.564491302 +0000 UTC m=+1105.308037553" watchObservedRunningTime="2025-12-10 09:52:22.576118071 +0000 UTC m=+1105.319664322" Dec 10 09:52:22 crc kubenswrapper[4715]: I1210 09:52:22.631461 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67f6cc7984-tshct" podStartSLOduration=26.070666914 podStartE2EDuration="26.631430807s" podCreationTimestamp="2025-12-10 09:51:56 +0000 UTC" firstStartedPulling="2025-12-10 09:52:20.316467017 +0000 UTC m=+1103.060013268" lastFinishedPulling="2025-12-10 09:52:20.87723091 +0000 UTC m=+1103.620777161" observedRunningTime="2025-12-10 09:52:22.596615061 +0000 UTC m=+1105.340161312" watchObservedRunningTime="2025-12-10 09:52:22.631430807 +0000 UTC m=+1105.374977068" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.574055 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85df69bf85-6cwpq" event={"ID":"0cf58c4f-3594-4b20-a3fb-313c0f25a844","Type":"ContainerStarted","Data":"4b905516d8ecb06adefe88bbdd56f23ff05d354ce4811997b0553f911e33e2d8"} Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.574502 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85df69bf85-6cwpq" event={"ID":"0cf58c4f-3594-4b20-a3fb-313c0f25a844","Type":"ContainerStarted","Data":"6b180850eee26d2760c865ac4631c6ec2997378fc8bd0d9520ab3b1ec6f6a84c"} Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.574519 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85df69bf85-6cwpq" event={"ID":"0cf58c4f-3594-4b20-a3fb-313c0f25a844","Type":"ContainerStarted","Data":"9d99b4602dca737ce57581d909b2632cacece949c76bac94ae734df97c6d1d22"} Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.575377 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.582324 4715 generic.go:334] "Generic (PLEG): container finished" podID="53507dfd-6f17-486f-8770-c073ec1b42a3" containerID="c51c3b816413cb9c1bd195a8e5ae598869eacdc3310e4b370b5d49c164de15ad" exitCode=0 Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.582389 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zq2xt" event={"ID":"53507dfd-6f17-486f-8770-c073ec1b42a3","Type":"ContainerDied","Data":"c51c3b816413cb9c1bd195a8e5ae598869eacdc3310e4b370b5d49c164de15ad"} Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.584639 4715 generic.go:334] "Generic (PLEG): container finished" podID="7c0209db-d227-4d43-a132-e2393ce0d5c3" containerID="c3a435ba5c7d59aec6834021bd7ea496b620189b0bcf79a4db85af3b4e6c81de" exitCode=0 Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.584672 4715 generic.go:334] "Generic (PLEG): container finished" podID="7c0209db-d227-4d43-a132-e2393ce0d5c3" containerID="a6a3751858a858b59e78c6e794614c1bb47d642d5cd6c2c664dee03ecd418eb9" exitCode=143 Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.584721 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7c0209db-d227-4d43-a132-e2393ce0d5c3","Type":"ContainerDied","Data":"c3a435ba5c7d59aec6834021bd7ea496b620189b0bcf79a4db85af3b4e6c81de"} Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.584761 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7c0209db-d227-4d43-a132-e2393ce0d5c3","Type":"ContainerDied","Data":"a6a3751858a858b59e78c6e794614c1bb47d642d5cd6c2c664dee03ecd418eb9"} Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.584783 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7c0209db-d227-4d43-a132-e2393ce0d5c3","Type":"ContainerDied","Data":"9f4f99ad8eb11aa447f1bce8ab9c0bbcc65324a07a2a8c0afec035773111102d"} Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.584796 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f4f99ad8eb11aa447f1bce8ab9c0bbcc65324a07a2a8c0afec035773111102d" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.586623 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.586671 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"207b4574-9a00-43c1-9d73-24e23343f35c","Type":"ContainerStarted","Data":"c37868688530d331997fbb4762beee580a5d5cba9d10c530844a74cea4b527c7"} Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.606689 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-85df69bf85-6cwpq" podStartSLOduration=2.6066695319999997 podStartE2EDuration="2.606669532s" podCreationTimestamp="2025-12-10 09:52:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:23.59457863 +0000 UTC m=+1106.338124881" watchObservedRunningTime="2025-12-10 09:52:23.606669532 +0000 UTC m=+1106.350215783" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.692412 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=21.691562035 podStartE2EDuration="21.691562035s" podCreationTimestamp="2025-12-10 09:52:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:23.66026926 +0000 UTC m=+1106.403815531" watchObservedRunningTime="2025-12-10 09:52:23.691562035 +0000 UTC m=+1106.435108286" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.724157 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-logs\") pod \"7c0209db-d227-4d43-a132-e2393ce0d5c3\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.724229 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-httpd-run\") pod \"7c0209db-d227-4d43-a132-e2393ce0d5c3\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.724297 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-combined-ca-bundle\") pod \"7c0209db-d227-4d43-a132-e2393ce0d5c3\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.724342 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-config-data\") pod \"7c0209db-d227-4d43-a132-e2393ce0d5c3\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.724416 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq4dg\" (UniqueName: \"kubernetes.io/projected/7c0209db-d227-4d43-a132-e2393ce0d5c3-kube-api-access-wq4dg\") pod \"7c0209db-d227-4d43-a132-e2393ce0d5c3\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.724476 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"7c0209db-d227-4d43-a132-e2393ce0d5c3\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.724517 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-scripts\") pod \"7c0209db-d227-4d43-a132-e2393ce0d5c3\" (UID: \"7c0209db-d227-4d43-a132-e2393ce0d5c3\") " Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.724715 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-logs" (OuterVolumeSpecName: "logs") pod "7c0209db-d227-4d43-a132-e2393ce0d5c3" (UID: "7c0209db-d227-4d43-a132-e2393ce0d5c3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.724844 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7c0209db-d227-4d43-a132-e2393ce0d5c3" (UID: "7c0209db-d227-4d43-a132-e2393ce0d5c3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.725585 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.725848 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c0209db-d227-4d43-a132-e2393ce0d5c3-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.738812 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "7c0209db-d227-4d43-a132-e2393ce0d5c3" (UID: "7c0209db-d227-4d43-a132-e2393ce0d5c3"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.738883 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c0209db-d227-4d43-a132-e2393ce0d5c3-kube-api-access-wq4dg" (OuterVolumeSpecName: "kube-api-access-wq4dg") pod "7c0209db-d227-4d43-a132-e2393ce0d5c3" (UID: "7c0209db-d227-4d43-a132-e2393ce0d5c3"). InnerVolumeSpecName "kube-api-access-wq4dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.760535 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-scripts" (OuterVolumeSpecName: "scripts") pod "7c0209db-d227-4d43-a132-e2393ce0d5c3" (UID: "7c0209db-d227-4d43-a132-e2393ce0d5c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.773115 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c0209db-d227-4d43-a132-e2393ce0d5c3" (UID: "7c0209db-d227-4d43-a132-e2393ce0d5c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.808108 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-config-data" (OuterVolumeSpecName: "config-data") pod "7c0209db-d227-4d43-a132-e2393ce0d5c3" (UID: "7c0209db-d227-4d43-a132-e2393ce0d5c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.827499 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq4dg\" (UniqueName: \"kubernetes.io/projected/7c0209db-d227-4d43-a132-e2393ce0d5c3-kube-api-access-wq4dg\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.827595 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.827609 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.827624 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.827634 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c0209db-d227-4d43-a132-e2393ce0d5c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.851697 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 10 09:52:23 crc kubenswrapper[4715]: I1210 09:52:23.934130 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.594701 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.650449 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.709604 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.737712 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:52:24 crc kubenswrapper[4715]: E1210 09:52:24.738575 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0209db-d227-4d43-a132-e2393ce0d5c3" containerName="glance-log" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.738607 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0209db-d227-4d43-a132-e2393ce0d5c3" containerName="glance-log" Dec 10 09:52:24 crc kubenswrapper[4715]: E1210 09:52:24.738619 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0209db-d227-4d43-a132-e2393ce0d5c3" containerName="glance-httpd" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.738627 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0209db-d227-4d43-a132-e2393ce0d5c3" containerName="glance-httpd" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.738881 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c0209db-d227-4d43-a132-e2393ce0d5c3" containerName="glance-httpd" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.738936 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c0209db-d227-4d43-a132-e2393ce0d5c3" containerName="glance-log" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.740621 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.743596 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.744425 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.747181 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.910115 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-logs\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.910388 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.910436 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.910472 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.910491 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.910536 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z8sm\" (UniqueName: \"kubernetes.io/projected/28610e7f-ab63-4b49-b9d5-73b17937f1b6-kube-api-access-2z8sm\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.910560 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:24 crc kubenswrapper[4715]: I1210 09:52:24.910645 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.012851 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z8sm\" (UniqueName: \"kubernetes.io/projected/28610e7f-ab63-4b49-b9d5-73b17937f1b6-kube-api-access-2z8sm\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.012901 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.013032 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.013104 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-logs\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.013151 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.013193 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.013227 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.013247 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.013694 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.015980 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-logs\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.016551 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.020057 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.020656 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.021195 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.022790 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.031570 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z8sm\" (UniqueName: \"kubernetes.io/projected/28610e7f-ab63-4b49-b9d5-73b17937f1b6-kube-api-access-2z8sm\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.056903 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.067752 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.175024 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zq2xt" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.317534 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-combined-ca-bundle\") pod \"53507dfd-6f17-486f-8770-c073ec1b42a3\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.317699 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsrcg\" (UniqueName: \"kubernetes.io/projected/53507dfd-6f17-486f-8770-c073ec1b42a3-kube-api-access-wsrcg\") pod \"53507dfd-6f17-486f-8770-c073ec1b42a3\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.317754 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-scripts\") pod \"53507dfd-6f17-486f-8770-c073ec1b42a3\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.317814 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-config-data\") pod \"53507dfd-6f17-486f-8770-c073ec1b42a3\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.317868 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53507dfd-6f17-486f-8770-c073ec1b42a3-logs\") pod \"53507dfd-6f17-486f-8770-c073ec1b42a3\" (UID: \"53507dfd-6f17-486f-8770-c073ec1b42a3\") " Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.318557 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53507dfd-6f17-486f-8770-c073ec1b42a3-logs" (OuterVolumeSpecName: "logs") pod "53507dfd-6f17-486f-8770-c073ec1b42a3" (UID: "53507dfd-6f17-486f-8770-c073ec1b42a3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.337104 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-scripts" (OuterVolumeSpecName: "scripts") pod "53507dfd-6f17-486f-8770-c073ec1b42a3" (UID: "53507dfd-6f17-486f-8770-c073ec1b42a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.340407 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53507dfd-6f17-486f-8770-c073ec1b42a3-kube-api-access-wsrcg" (OuterVolumeSpecName: "kube-api-access-wsrcg") pod "53507dfd-6f17-486f-8770-c073ec1b42a3" (UID: "53507dfd-6f17-486f-8770-c073ec1b42a3"). InnerVolumeSpecName "kube-api-access-wsrcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.347149 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53507dfd-6f17-486f-8770-c073ec1b42a3" (UID: "53507dfd-6f17-486f-8770-c073ec1b42a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.370646 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-config-data" (OuterVolumeSpecName: "config-data") pod "53507dfd-6f17-486f-8770-c073ec1b42a3" (UID: "53507dfd-6f17-486f-8770-c073ec1b42a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.421947 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsrcg\" (UniqueName: \"kubernetes.io/projected/53507dfd-6f17-486f-8770-c073ec1b42a3-kube-api-access-wsrcg\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.422297 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.422311 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.422327 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53507dfd-6f17-486f-8770-c073ec1b42a3-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.422339 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53507dfd-6f17-486f-8770-c073ec1b42a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.607835 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zq2xt" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.617578 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c0209db-d227-4d43-a132-e2393ce0d5c3" path="/var/lib/kubelet/pods/7c0209db-d227-4d43-a132-e2393ce0d5c3/volumes" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.618770 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zq2xt" event={"ID":"53507dfd-6f17-486f-8770-c073ec1b42a3","Type":"ContainerDied","Data":"2f3fc708d59ff69d9105a8e945f2864324b8786ef2108b561cd706bb6ecdfb86"} Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.618804 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f3fc708d59ff69d9105a8e945f2864324b8786ef2108b561cd706bb6ecdfb86" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.789305 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.830023 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5777d57b56-6lbcp"] Dec 10 09:52:25 crc kubenswrapper[4715]: E1210 09:52:25.830509 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53507dfd-6f17-486f-8770-c073ec1b42a3" containerName="placement-db-sync" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.830526 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="53507dfd-6f17-486f-8770-c073ec1b42a3" containerName="placement-db-sync" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.830774 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="53507dfd-6f17-486f-8770-c073ec1b42a3" containerName="placement-db-sync" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.832405 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.835520 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.835766 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.835900 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-fnshj" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.836052 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.836234 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.856035 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5777d57b56-6lbcp"] Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.932572 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-internal-tls-certs\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.932634 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-combined-ca-bundle\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.932708 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-scripts\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.932735 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21593c1a-bc1a-4460-9ef9-3eef12261355-logs\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.932764 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-public-tls-certs\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.932812 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-config-data\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:25 crc kubenswrapper[4715]: I1210 09:52:25.932929 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt872\" (UniqueName: \"kubernetes.io/projected/21593c1a-bc1a-4460-9ef9-3eef12261355-kube-api-access-lt872\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.037425 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-internal-tls-certs\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.037497 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-combined-ca-bundle\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.037549 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-scripts\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.037582 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21593c1a-bc1a-4460-9ef9-3eef12261355-logs\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.037614 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-public-tls-certs\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.037679 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-config-data\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.037724 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt872\" (UniqueName: \"kubernetes.io/projected/21593c1a-bc1a-4460-9ef9-3eef12261355-kube-api-access-lt872\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.040154 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21593c1a-bc1a-4460-9ef9-3eef12261355-logs\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.046083 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-config-data\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.046766 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-public-tls-certs\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.047342 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-internal-tls-certs\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.051947 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-combined-ca-bundle\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.053433 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21593c1a-bc1a-4460-9ef9-3eef12261355-scripts\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.063114 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt872\" (UniqueName: \"kubernetes.io/projected/21593c1a-bc1a-4460-9ef9-3eef12261355-kube-api-access-lt872\") pod \"placement-5777d57b56-6lbcp\" (UID: \"21593c1a-bc1a-4460-9ef9-3eef12261355\") " pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.183482 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.648867 4715 generic.go:334] "Generic (PLEG): container finished" podID="27133872-3a45-45a6-9d91-f645c0c9b264" containerID="9ef9793a46d67594228c5ad70a237a786149e60b2da549a0aad6cb2b5bcc7930" exitCode=0 Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.649092 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x2v2z" event={"ID":"27133872-3a45-45a6-9d91-f645c0c9b264","Type":"ContainerDied","Data":"9ef9793a46d67594228c5ad70a237a786149e60b2da549a0aad6cb2b5bcc7930"} Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.655401 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"28610e7f-ab63-4b49-b9d5-73b17937f1b6","Type":"ContainerStarted","Data":"8ff58dd4c00af545e2be2e5ebc3e656b8e492796b8d42a11c5a8b4dbbb62373b"} Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.655434 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"28610e7f-ab63-4b49-b9d5-73b17937f1b6","Type":"ContainerStarted","Data":"c42ac83d795f3d8d04ef9effa327a8d30a5066d09fa57a026b9bbb64223a3f89"} Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.683073 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5777d57b56-6lbcp"] Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.861082 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:52:26 crc kubenswrapper[4715]: I1210 09:52:26.861998 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:52:27 crc kubenswrapper[4715]: I1210 09:52:27.087533 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:52:27 crc kubenswrapper[4715]: I1210 09:52:27.087662 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:52:28 crc kubenswrapper[4715]: I1210 09:52:28.966162 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:29 crc kubenswrapper[4715]: I1210 09:52:29.053799 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-8jbht"] Dec 10 09:52:29 crc kubenswrapper[4715]: I1210 09:52:29.054043 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" podUID="0c33dcac-9cd8-41e9-b4b2-ba521a728286" containerName="dnsmasq-dns" containerID="cri-o://8cabc5a59d136bdb08c1067fda57027192f115e00ce85b89dd35bcc8c305f28c" gracePeriod=10 Dec 10 09:52:29 crc kubenswrapper[4715]: I1210 09:52:29.686740 4715 generic.go:334] "Generic (PLEG): container finished" podID="0c33dcac-9cd8-41e9-b4b2-ba521a728286" containerID="8cabc5a59d136bdb08c1067fda57027192f115e00ce85b89dd35bcc8c305f28c" exitCode=0 Dec 10 09:52:29 crc kubenswrapper[4715]: I1210 09:52:29.686787 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" event={"ID":"0c33dcac-9cd8-41e9-b4b2-ba521a728286","Type":"ContainerDied","Data":"8cabc5a59d136bdb08c1067fda57027192f115e00ce85b89dd35bcc8c305f28c"} Dec 10 09:52:31 crc kubenswrapper[4715]: I1210 09:52:31.955778 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" podUID="0c33dcac-9cd8-41e9-b4b2-ba521a728286" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: connect: connection refused" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.178888 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.248517 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.354869 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-credential-keys\") pod \"27133872-3a45-45a6-9d91-f645c0c9b264\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.354963 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-swift-storage-0\") pod \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.355004 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-fernet-keys\") pod \"27133872-3a45-45a6-9d91-f645c0c9b264\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.355026 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-config-data\") pod \"27133872-3a45-45a6-9d91-f645c0c9b264\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.355120 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-config\") pod \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.355145 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-combined-ca-bundle\") pod \"27133872-3a45-45a6-9d91-f645c0c9b264\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.355210 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-svc\") pod \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.355237 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-nb\") pod \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.355269 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-scripts\") pod \"27133872-3a45-45a6-9d91-f645c0c9b264\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.355286 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkblm\" (UniqueName: \"kubernetes.io/projected/0c33dcac-9cd8-41e9-b4b2-ba521a728286-kube-api-access-gkblm\") pod \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.355320 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjrn5\" (UniqueName: \"kubernetes.io/projected/27133872-3a45-45a6-9d91-f645c0c9b264-kube-api-access-bjrn5\") pod \"27133872-3a45-45a6-9d91-f645c0c9b264\" (UID: \"27133872-3a45-45a6-9d91-f645c0c9b264\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.355358 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-sb\") pod \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\" (UID: \"0c33dcac-9cd8-41e9-b4b2-ba521a728286\") " Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.364261 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27133872-3a45-45a6-9d91-f645c0c9b264-kube-api-access-bjrn5" (OuterVolumeSpecName: "kube-api-access-bjrn5") pod "27133872-3a45-45a6-9d91-f645c0c9b264" (UID: "27133872-3a45-45a6-9d91-f645c0c9b264"). InnerVolumeSpecName "kube-api-access-bjrn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.364640 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c33dcac-9cd8-41e9-b4b2-ba521a728286-kube-api-access-gkblm" (OuterVolumeSpecName: "kube-api-access-gkblm") pod "0c33dcac-9cd8-41e9-b4b2-ba521a728286" (UID: "0c33dcac-9cd8-41e9-b4b2-ba521a728286"). InnerVolumeSpecName "kube-api-access-gkblm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.368382 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "27133872-3a45-45a6-9d91-f645c0c9b264" (UID: "27133872-3a45-45a6-9d91-f645c0c9b264"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.368430 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-scripts" (OuterVolumeSpecName: "scripts") pod "27133872-3a45-45a6-9d91-f645c0c9b264" (UID: "27133872-3a45-45a6-9d91-f645c0c9b264"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.368547 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "27133872-3a45-45a6-9d91-f645c0c9b264" (UID: "27133872-3a45-45a6-9d91-f645c0c9b264"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.391212 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-config-data" (OuterVolumeSpecName: "config-data") pod "27133872-3a45-45a6-9d91-f645c0c9b264" (UID: "27133872-3a45-45a6-9d91-f645c0c9b264"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.407672 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0c33dcac-9cd8-41e9-b4b2-ba521a728286" (UID: "0c33dcac-9cd8-41e9-b4b2-ba521a728286"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.420575 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-config" (OuterVolumeSpecName: "config") pod "0c33dcac-9cd8-41e9-b4b2-ba521a728286" (UID: "0c33dcac-9cd8-41e9-b4b2-ba521a728286"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.422977 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0c33dcac-9cd8-41e9-b4b2-ba521a728286" (UID: "0c33dcac-9cd8-41e9-b4b2-ba521a728286"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.427321 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27133872-3a45-45a6-9d91-f645c0c9b264" (UID: "27133872-3a45-45a6-9d91-f645c0c9b264"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.438386 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0c33dcac-9cd8-41e9-b4b2-ba521a728286" (UID: "0c33dcac-9cd8-41e9-b4b2-ba521a728286"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.440540 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0c33dcac-9cd8-41e9-b4b2-ba521a728286" (UID: "0c33dcac-9cd8-41e9-b4b2-ba521a728286"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.457756 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.457798 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.457812 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkblm\" (UniqueName: \"kubernetes.io/projected/0c33dcac-9cd8-41e9-b4b2-ba521a728286-kube-api-access-gkblm\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.457825 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjrn5\" (UniqueName: \"kubernetes.io/projected/27133872-3a45-45a6-9d91-f645c0c9b264-kube-api-access-bjrn5\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.457833 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.457840 4715 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.457997 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.458014 4715 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.458024 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.458036 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.458047 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27133872-3a45-45a6-9d91-f645c0c9b264-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.458080 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c33dcac-9cd8-41e9-b4b2-ba521a728286-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.737898 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" event={"ID":"0c33dcac-9cd8-41e9-b4b2-ba521a728286","Type":"ContainerDied","Data":"63c22a057e78cb913772e8849661d111d991d6f7593f2b1ca61f99f4b9e51b87"} Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.737977 4715 scope.go:117] "RemoveContainer" containerID="8cabc5a59d136bdb08c1067fda57027192f115e00ce85b89dd35bcc8c305f28c" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.738150 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-8jbht" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.742357 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2167d98b-0365-421e-9d3d-6ff62f7bfb61","Type":"ContainerStarted","Data":"d6790cb6e37145805338959fadfadc2c40fef0faaee91d9db225c0f632e72da0"} Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.745490 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5777d57b56-6lbcp" event={"ID":"21593c1a-bc1a-4460-9ef9-3eef12261355","Type":"ContainerStarted","Data":"fb9f02eb1f8f4b6c48d62a691b273c8fcab82c55b2db31c9ad9a884e7b343ecb"} Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.745554 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5777d57b56-6lbcp" event={"ID":"21593c1a-bc1a-4460-9ef9-3eef12261355","Type":"ContainerStarted","Data":"a1b23b81a677ad873d3d328170368e835a2cf9db104d3e745618bd74187e1936"} Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.745570 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5777d57b56-6lbcp" event={"ID":"21593c1a-bc1a-4460-9ef9-3eef12261355","Type":"ContainerStarted","Data":"c97efa8d522b7f123d0451386c9a13b4200c334502d0145750878a98adaaa039"} Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.746377 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.746590 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.750310 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x2v2z" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.751006 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x2v2z" event={"ID":"27133872-3a45-45a6-9d91-f645c0c9b264","Type":"ContainerDied","Data":"c38a582938016190564b921cc60a17793d2e4debe2d26c5cf1dfe6afbbbcc83c"} Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.751050 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c38a582938016190564b921cc60a17793d2e4debe2d26c5cf1dfe6afbbbcc83c" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.754539 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.754581 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.754598 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.755714 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.772664 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"28610e7f-ab63-4b49-b9d5-73b17937f1b6","Type":"ContainerStarted","Data":"51e47c5a22fe27694a01c0e2b0c68edc82cd7c5c052747e4d8a03a8164dbf545"} Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.773373 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5777d57b56-6lbcp" podStartSLOduration=7.773351605 podStartE2EDuration="7.773351605s" podCreationTimestamp="2025-12-10 09:52:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:32.773329324 +0000 UTC m=+1115.516875575" watchObservedRunningTime="2025-12-10 09:52:32.773351605 +0000 UTC m=+1115.516897856" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.807805 4715 scope.go:117] "RemoveContainer" containerID="9abec03902435ff97b9dc6dfc48747fe8fb38d18f71a7cb6cccbd9c7b493001c" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.809674 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.809654002 podStartE2EDuration="8.809654002s" podCreationTimestamp="2025-12-10 09:52:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:32.808824239 +0000 UTC m=+1115.552370500" watchObservedRunningTime="2025-12-10 09:52:32.809654002 +0000 UTC m=+1115.553200253" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.830288 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.839150 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-8jbht"] Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.848303 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-8jbht"] Dec 10 09:52:32 crc kubenswrapper[4715]: I1210 09:52:32.851017 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.308574 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-d9c5c96bf-n94nb"] Dec 10 09:52:33 crc kubenswrapper[4715]: E1210 09:52:33.308926 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c33dcac-9cd8-41e9-b4b2-ba521a728286" containerName="init" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.308937 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c33dcac-9cd8-41e9-b4b2-ba521a728286" containerName="init" Dec 10 09:52:33 crc kubenswrapper[4715]: E1210 09:52:33.308965 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c33dcac-9cd8-41e9-b4b2-ba521a728286" containerName="dnsmasq-dns" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.308971 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c33dcac-9cd8-41e9-b4b2-ba521a728286" containerName="dnsmasq-dns" Dec 10 09:52:33 crc kubenswrapper[4715]: E1210 09:52:33.308989 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27133872-3a45-45a6-9d91-f645c0c9b264" containerName="keystone-bootstrap" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.308995 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="27133872-3a45-45a6-9d91-f645c0c9b264" containerName="keystone-bootstrap" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.309144 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="27133872-3a45-45a6-9d91-f645c0c9b264" containerName="keystone-bootstrap" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.309154 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c33dcac-9cd8-41e9-b4b2-ba521a728286" containerName="dnsmasq-dns" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.309735 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.314359 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.314420 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-csxkd" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.314513 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.314735 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.314891 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.315092 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.326830 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d9c5c96bf-n94nb"] Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.492091 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-internal-tls-certs\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.492181 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-credential-keys\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.492293 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-scripts\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.492340 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-config-data\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.492434 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-public-tls-certs\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.492462 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhhb4\" (UniqueName: \"kubernetes.io/projected/946c4b3b-eb91-4d39-a7ef-88d25e23599e-kube-api-access-zhhb4\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.492491 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-combined-ca-bundle\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.492544 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-fernet-keys\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.596026 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-public-tls-certs\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.596090 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhhb4\" (UniqueName: \"kubernetes.io/projected/946c4b3b-eb91-4d39-a7ef-88d25e23599e-kube-api-access-zhhb4\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.596130 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-combined-ca-bundle\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.596157 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-fernet-keys\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.596214 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-internal-tls-certs\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.596251 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-credential-keys\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.596402 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-scripts\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.596457 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-config-data\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.600371 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-public-tls-certs\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.601510 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-credential-keys\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.601641 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-internal-tls-certs\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.603497 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-combined-ca-bundle\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.604785 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-config-data\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.612616 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-fernet-keys\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.617365 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/946c4b3b-eb91-4d39-a7ef-88d25e23599e-scripts\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.622698 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhhb4\" (UniqueName: \"kubernetes.io/projected/946c4b3b-eb91-4d39-a7ef-88d25e23599e-kube-api-access-zhhb4\") pod \"keystone-d9c5c96bf-n94nb\" (UID: \"946c4b3b-eb91-4d39-a7ef-88d25e23599e\") " pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.625440 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c33dcac-9cd8-41e9-b4b2-ba521a728286" path="/var/lib/kubelet/pods/0c33dcac-9cd8-41e9-b4b2-ba521a728286/volumes" Dec 10 09:52:33 crc kubenswrapper[4715]: I1210 09:52:33.637483 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:34 crc kubenswrapper[4715]: I1210 09:52:34.112696 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d9c5c96bf-n94nb"] Dec 10 09:52:34 crc kubenswrapper[4715]: I1210 09:52:34.804314 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d9c5c96bf-n94nb" event={"ID":"946c4b3b-eb91-4d39-a7ef-88d25e23599e","Type":"ContainerStarted","Data":"e84ef96cf3edb48c6aebc7110b34a5b088d28ab5bfb458c9262aef6bb7ec5a9c"} Dec 10 09:52:34 crc kubenswrapper[4715]: I1210 09:52:34.804642 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d9c5c96bf-n94nb" event={"ID":"946c4b3b-eb91-4d39-a7ef-88d25e23599e","Type":"ContainerStarted","Data":"8916d29b15ad0c1d710d24ef1d3e0dd5d477f9354f1c57ffeb208678e8fb7b9d"} Dec 10 09:52:34 crc kubenswrapper[4715]: I1210 09:52:34.807382 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gzr48" event={"ID":"66856339-f4ab-4102-93a8-adc952877ccd","Type":"ContainerStarted","Data":"069f829a7a1f67a425ce726262e67f467cf00ece44d7379725c41a6b91c5c453"} Dec 10 09:52:34 crc kubenswrapper[4715]: I1210 09:52:34.826495 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-gzr48" podStartSLOduration=4.379313655 podStartE2EDuration="49.826480952s" podCreationTimestamp="2025-12-10 09:51:45 +0000 UTC" firstStartedPulling="2025-12-10 09:51:48.087715945 +0000 UTC m=+1070.831262196" lastFinishedPulling="2025-12-10 09:52:33.534883242 +0000 UTC m=+1116.278429493" observedRunningTime="2025-12-10 09:52:34.81969705 +0000 UTC m=+1117.563243301" watchObservedRunningTime="2025-12-10 09:52:34.826480952 +0000 UTC m=+1117.570027203" Dec 10 09:52:35 crc kubenswrapper[4715]: I1210 09:52:35.068501 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 09:52:35 crc kubenswrapper[4715]: I1210 09:52:35.068832 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 09:52:35 crc kubenswrapper[4715]: I1210 09:52:35.112690 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 09:52:35 crc kubenswrapper[4715]: I1210 09:52:35.119626 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 09:52:35 crc kubenswrapper[4715]: I1210 09:52:35.823449 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 09:52:35 crc kubenswrapper[4715]: I1210 09:52:35.823498 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 09:52:36 crc kubenswrapper[4715]: I1210 09:52:36.308784 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:36 crc kubenswrapper[4715]: I1210 09:52:36.309396 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:52:36 crc kubenswrapper[4715]: I1210 09:52:36.494535 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 09:52:36 crc kubenswrapper[4715]: I1210 09:52:36.832495 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:52:36 crc kubenswrapper[4715]: I1210 09:52:36.856690 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-d9c5c96bf-n94nb" podStartSLOduration=3.856674652 podStartE2EDuration="3.856674652s" podCreationTimestamp="2025-12-10 09:52:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:36.850050644 +0000 UTC m=+1119.593596895" watchObservedRunningTime="2025-12-10 09:52:36.856674652 +0000 UTC m=+1119.600220903" Dec 10 09:52:36 crc kubenswrapper[4715]: I1210 09:52:36.862807 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-67f6cc7984-tshct" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 10 09:52:37 crc kubenswrapper[4715]: I1210 09:52:37.089500 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-857b4dfcd4-wnlbz" podUID="4214260f-423f-47fb-b062-c752f3519175" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 10 09:52:37 crc kubenswrapper[4715]: I1210 09:52:37.860045 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-28hsp" event={"ID":"8bf72f3c-8033-4018-b330-72fa2774f402","Type":"ContainerStarted","Data":"1a48560f27a8b6743404f92c71729c3d1067c20585ceae502096c59d41300cb0"} Dec 10 09:52:37 crc kubenswrapper[4715]: I1210 09:52:37.876633 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-28hsp" podStartSLOduration=4.926139105 podStartE2EDuration="52.876618102s" podCreationTimestamp="2025-12-10 09:51:45 +0000 UTC" firstStartedPulling="2025-12-10 09:51:48.103459431 +0000 UTC m=+1070.847005682" lastFinishedPulling="2025-12-10 09:52:36.053938418 +0000 UTC m=+1118.797484679" observedRunningTime="2025-12-10 09:52:37.873331269 +0000 UTC m=+1120.616877520" watchObservedRunningTime="2025-12-10 09:52:37.876618102 +0000 UTC m=+1120.620164353" Dec 10 09:52:38 crc kubenswrapper[4715]: I1210 09:52:38.412547 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 09:52:38 crc kubenswrapper[4715]: I1210 09:52:38.874516 4715 generic.go:334] "Generic (PLEG): container finished" podID="66856339-f4ab-4102-93a8-adc952877ccd" containerID="069f829a7a1f67a425ce726262e67f467cf00ece44d7379725c41a6b91c5c453" exitCode=0 Dec 10 09:52:38 crc kubenswrapper[4715]: I1210 09:52:38.874593 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gzr48" event={"ID":"66856339-f4ab-4102-93a8-adc952877ccd","Type":"ContainerDied","Data":"069f829a7a1f67a425ce726262e67f467cf00ece44d7379725c41a6b91c5c453"} Dec 10 09:52:40 crc kubenswrapper[4715]: I1210 09:52:40.342402 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.381887 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gzr48" Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.465003 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz8v7\" (UniqueName: \"kubernetes.io/projected/66856339-f4ab-4102-93a8-adc952877ccd-kube-api-access-hz8v7\") pod \"66856339-f4ab-4102-93a8-adc952877ccd\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.465075 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-db-sync-config-data\") pod \"66856339-f4ab-4102-93a8-adc952877ccd\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.465173 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-combined-ca-bundle\") pod \"66856339-f4ab-4102-93a8-adc952877ccd\" (UID: \"66856339-f4ab-4102-93a8-adc952877ccd\") " Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.472950 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "66856339-f4ab-4102-93a8-adc952877ccd" (UID: "66856339-f4ab-4102-93a8-adc952877ccd"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.486819 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66856339-f4ab-4102-93a8-adc952877ccd-kube-api-access-hz8v7" (OuterVolumeSpecName: "kube-api-access-hz8v7") pod "66856339-f4ab-4102-93a8-adc952877ccd" (UID: "66856339-f4ab-4102-93a8-adc952877ccd"). InnerVolumeSpecName "kube-api-access-hz8v7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.496641 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66856339-f4ab-4102-93a8-adc952877ccd" (UID: "66856339-f4ab-4102-93a8-adc952877ccd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.567866 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz8v7\" (UniqueName: \"kubernetes.io/projected/66856339-f4ab-4102-93a8-adc952877ccd-kube-api-access-hz8v7\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.567901 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.568000 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66856339-f4ab-4102-93a8-adc952877ccd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.925199 4715 generic.go:334] "Generic (PLEG): container finished" podID="8bf72f3c-8033-4018-b330-72fa2774f402" containerID="1a48560f27a8b6743404f92c71729c3d1067c20585ceae502096c59d41300cb0" exitCode=0 Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.925282 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-28hsp" event={"ID":"8bf72f3c-8033-4018-b330-72fa2774f402","Type":"ContainerDied","Data":"1a48560f27a8b6743404f92c71729c3d1067c20585ceae502096c59d41300cb0"} Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.930645 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-gzr48" event={"ID":"66856339-f4ab-4102-93a8-adc952877ccd","Type":"ContainerDied","Data":"cb728b42ad1d452c5e1bc5c830dc1a5d39e4abe4cca259bed26503c9eeed0eec"} Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.930687 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb728b42ad1d452c5e1bc5c830dc1a5d39e4abe4cca259bed26503c9eeed0eec" Dec 10 09:52:41 crc kubenswrapper[4715]: I1210 09:52:41.930750 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-gzr48" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.664976 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6d588649f8-m9jx2"] Dec 10 09:52:42 crc kubenswrapper[4715]: E1210 09:52:42.666749 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66856339-f4ab-4102-93a8-adc952877ccd" containerName="barbican-db-sync" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.666839 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="66856339-f4ab-4102-93a8-adc952877ccd" containerName="barbican-db-sync" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.667159 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="66856339-f4ab-4102-93a8-adc952877ccd" containerName="barbican-db-sync" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.668444 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.679142 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.681360 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.685744 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7fdl\" (UniqueName: \"kubernetes.io/projected/73357509-8bf3-408e-9c35-ee6267593be1-kube-api-access-v7fdl\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.685848 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73357509-8bf3-408e-9c35-ee6267593be1-logs\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.685940 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73357509-8bf3-408e-9c35-ee6267593be1-config-data\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.685993 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73357509-8bf3-408e-9c35-ee6267593be1-config-data-custom\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.686015 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73357509-8bf3-408e-9c35-ee6267593be1-combined-ca-bundle\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.688827 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9fdrx" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.715324 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6d588649f8-m9jx2"] Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.744046 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5c964fbd47-67rln"] Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.745820 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.757278 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.767032 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c964fbd47-67rln"] Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.790423 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvrrd\" (UniqueName: \"kubernetes.io/projected/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-kube-api-access-zvrrd\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.790491 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73357509-8bf3-408e-9c35-ee6267593be1-logs\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.790647 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-combined-ca-bundle\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.790748 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-config-data-custom\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.790880 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73357509-8bf3-408e-9c35-ee6267593be1-config-data\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.790959 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73357509-8bf3-408e-9c35-ee6267593be1-logs\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.790982 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73357509-8bf3-408e-9c35-ee6267593be1-config-data-custom\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.791016 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73357509-8bf3-408e-9c35-ee6267593be1-combined-ca-bundle\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.791175 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-logs\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.791199 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7fdl\" (UniqueName: \"kubernetes.io/projected/73357509-8bf3-408e-9c35-ee6267593be1-kube-api-access-v7fdl\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.791238 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-config-data\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.797151 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73357509-8bf3-408e-9c35-ee6267593be1-config-data-custom\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.802765 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73357509-8bf3-408e-9c35-ee6267593be1-combined-ca-bundle\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.805009 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73357509-8bf3-408e-9c35-ee6267593be1-config-data\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.884481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7fdl\" (UniqueName: \"kubernetes.io/projected/73357509-8bf3-408e-9c35-ee6267593be1-kube-api-access-v7fdl\") pod \"barbican-keystone-listener-6d588649f8-m9jx2\" (UID: \"73357509-8bf3-408e-9c35-ee6267593be1\") " pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.893350 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-combined-ca-bundle\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.893418 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-config-data-custom\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.893548 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-logs\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.893569 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-config-data\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.893633 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvrrd\" (UniqueName: \"kubernetes.io/projected/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-kube-api-access-zvrrd\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.895079 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-logs\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.912601 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-combined-ca-bundle\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.913174 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-config-data\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.920583 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-config-data-custom\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.930046 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8lgxc"] Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.931792 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.938975 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvrrd\" (UniqueName: \"kubernetes.io/projected/b46f5949-5bfa-4cfd-b70d-4d225ea12d6b-kube-api-access-zvrrd\") pod \"barbican-worker-5c964fbd47-67rln\" (UID: \"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b\") " pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.949938 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="ceilometer-central-agent" containerID="cri-o://382c1037c3c28ee70ddc99fb57d441e603a2485e73eef3c7410495d5fb30133b" gracePeriod=30 Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.950101 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2167d98b-0365-421e-9d3d-6ff62f7bfb61","Type":"ContainerStarted","Data":"5bf233fe7cd9fee035d356ea9cfed7ecbd952a02f2fe9e5efc4c25f25096bb7d"} Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.950158 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.950160 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="proxy-httpd" containerID="cri-o://5bf233fe7cd9fee035d356ea9cfed7ecbd952a02f2fe9e5efc4c25f25096bb7d" gracePeriod=30 Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.950193 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="ceilometer-notification-agent" containerID="cri-o://d6a23411f595071e43d6ed0df1b67d55411b62de11ff04597a980928622d1e76" gracePeriod=30 Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.950216 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="sg-core" containerID="cri-o://d6790cb6e37145805338959fadfadc2c40fef0faaee91d9db225c0f632e72da0" gracePeriod=30 Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.982584 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8lgxc"] Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.990107 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.997174 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.997253 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-config\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.997308 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-svc\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.997343 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.997365 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppxmz\" (UniqueName: \"kubernetes.io/projected/8029fef4-8c21-42fd-8c79-7d76f6919d6e-kube-api-access-ppxmz\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:42 crc kubenswrapper[4715]: I1210 09:52:42.997412 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.022992 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c86fc9d-jwnlv"] Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.024540 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.027884 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.235044743 podStartE2EDuration="58.027865339s" podCreationTimestamp="2025-12-10 09:51:45 +0000 UTC" firstStartedPulling="2025-12-10 09:51:48.087389326 +0000 UTC m=+1070.830935577" lastFinishedPulling="2025-12-10 09:52:41.880209922 +0000 UTC m=+1124.623756173" observedRunningTime="2025-12-10 09:52:43.014498531 +0000 UTC m=+1125.758044782" watchObservedRunningTime="2025-12-10 09:52:43.027865339 +0000 UTC m=+1125.771411580" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.028111 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.073525 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c964fbd47-67rln" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.084500 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c86fc9d-jwnlv"] Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099051 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-logs\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099386 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-svc\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099430 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099458 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099496 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppxmz\" (UniqueName: \"kubernetes.io/projected/8029fef4-8c21-42fd-8c79-7d76f6919d6e-kube-api-access-ppxmz\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099533 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-combined-ca-bundle\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099581 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099633 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkswn\" (UniqueName: \"kubernetes.io/projected/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-kube-api-access-jkswn\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099666 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099703 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data-custom\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.099753 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-config\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.100813 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-config\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.104674 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.105709 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-svc\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.106143 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.113385 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.131852 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppxmz\" (UniqueName: \"kubernetes.io/projected/8029fef4-8c21-42fd-8c79-7d76f6919d6e-kube-api-access-ppxmz\") pod \"dnsmasq-dns-85ff748b95-8lgxc\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.200875 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.200944 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-combined-ca-bundle\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.200997 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkswn\" (UniqueName: \"kubernetes.io/projected/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-kube-api-access-jkswn\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.201029 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data-custom\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.201090 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-logs\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.201492 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-logs\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.204992 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-combined-ca-bundle\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.205882 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.206098 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data-custom\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.224343 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkswn\" (UniqueName: \"kubernetes.io/projected/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-kube-api-access-jkswn\") pod \"barbican-api-6c86fc9d-jwnlv\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.346349 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.364205 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.580199 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-28hsp" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.731965 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6d588649f8-m9jx2"] Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.749976 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-scripts\") pod \"8bf72f3c-8033-4018-b330-72fa2774f402\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.750035 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bf72f3c-8033-4018-b330-72fa2774f402-etc-machine-id\") pod \"8bf72f3c-8033-4018-b330-72fa2774f402\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.750090 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-combined-ca-bundle\") pod \"8bf72f3c-8033-4018-b330-72fa2774f402\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.750158 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-db-sync-config-data\") pod \"8bf72f3c-8033-4018-b330-72fa2774f402\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.750278 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-config-data\") pod \"8bf72f3c-8033-4018-b330-72fa2774f402\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.750302 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwm8t\" (UniqueName: \"kubernetes.io/projected/8bf72f3c-8033-4018-b330-72fa2774f402-kube-api-access-dwm8t\") pod \"8bf72f3c-8033-4018-b330-72fa2774f402\" (UID: \"8bf72f3c-8033-4018-b330-72fa2774f402\") " Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.751080 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8bf72f3c-8033-4018-b330-72fa2774f402-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8bf72f3c-8033-4018-b330-72fa2774f402" (UID: "8bf72f3c-8033-4018-b330-72fa2774f402"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.755492 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-scripts" (OuterVolumeSpecName: "scripts") pod "8bf72f3c-8033-4018-b330-72fa2774f402" (UID: "8bf72f3c-8033-4018-b330-72fa2774f402"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.755605 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8bf72f3c-8033-4018-b330-72fa2774f402" (UID: "8bf72f3c-8033-4018-b330-72fa2774f402"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.756300 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf72f3c-8033-4018-b330-72fa2774f402-kube-api-access-dwm8t" (OuterVolumeSpecName: "kube-api-access-dwm8t") pod "8bf72f3c-8033-4018-b330-72fa2774f402" (UID: "8bf72f3c-8033-4018-b330-72fa2774f402"). InnerVolumeSpecName "kube-api-access-dwm8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.792138 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bf72f3c-8033-4018-b330-72fa2774f402" (UID: "8bf72f3c-8033-4018-b330-72fa2774f402"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.807587 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c964fbd47-67rln"] Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.828015 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-config-data" (OuterVolumeSpecName: "config-data") pod "8bf72f3c-8033-4018-b330-72fa2774f402" (UID: "8bf72f3c-8033-4018-b330-72fa2774f402"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.852218 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.852249 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwm8t\" (UniqueName: \"kubernetes.io/projected/8bf72f3c-8033-4018-b330-72fa2774f402-kube-api-access-dwm8t\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.852263 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.852273 4715 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bf72f3c-8033-4018-b330-72fa2774f402-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.852286 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.852297 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8bf72f3c-8033-4018-b330-72fa2774f402-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.980463 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-28hsp" event={"ID":"8bf72f3c-8033-4018-b330-72fa2774f402","Type":"ContainerDied","Data":"cd889dd567a1b1de84aff2de0dcc0debb5051bf39cfffbdfb6abd287785cc536"} Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.980610 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-28hsp" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.980676 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd889dd567a1b1de84aff2de0dcc0debb5051bf39cfffbdfb6abd287785cc536" Dec 10 09:52:43 crc kubenswrapper[4715]: I1210 09:52:43.993092 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" event={"ID":"73357509-8bf3-408e-9c35-ee6267593be1","Type":"ContainerStarted","Data":"2bcbd22174b077ece77f36f8562fc5bde07f98d4e52228f3263ada12b5fd18dc"} Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.006127 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8lgxc"] Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.007749 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c964fbd47-67rln" event={"ID":"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b","Type":"ContainerStarted","Data":"662ad31487f8b93d901d29557a40fa5e73b9fb45a87446c062dafa499b92236d"} Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.019640 4715 generic.go:334] "Generic (PLEG): container finished" podID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerID="5bf233fe7cd9fee035d356ea9cfed7ecbd952a02f2fe9e5efc4c25f25096bb7d" exitCode=0 Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.019674 4715 generic.go:334] "Generic (PLEG): container finished" podID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerID="d6790cb6e37145805338959fadfadc2c40fef0faaee91d9db225c0f632e72da0" exitCode=2 Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.019683 4715 generic.go:334] "Generic (PLEG): container finished" podID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerID="382c1037c3c28ee70ddc99fb57d441e603a2485e73eef3c7410495d5fb30133b" exitCode=0 Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.019703 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2167d98b-0365-421e-9d3d-6ff62f7bfb61","Type":"ContainerDied","Data":"5bf233fe7cd9fee035d356ea9cfed7ecbd952a02f2fe9e5efc4c25f25096bb7d"} Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.019729 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2167d98b-0365-421e-9d3d-6ff62f7bfb61","Type":"ContainerDied","Data":"d6790cb6e37145805338959fadfadc2c40fef0faaee91d9db225c0f632e72da0"} Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.019738 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2167d98b-0365-421e-9d3d-6ff62f7bfb61","Type":"ContainerDied","Data":"382c1037c3c28ee70ddc99fb57d441e603a2485e73eef3c7410495d5fb30133b"} Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.164654 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c86fc9d-jwnlv"] Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.299631 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 09:52:44 crc kubenswrapper[4715]: E1210 09:52:44.300114 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf72f3c-8033-4018-b330-72fa2774f402" containerName="cinder-db-sync" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.300126 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf72f3c-8033-4018-b330-72fa2774f402" containerName="cinder-db-sync" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.300360 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf72f3c-8033-4018-b330-72fa2774f402" containerName="cinder-db-sync" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.301610 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.305390 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.305560 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.305656 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.305754 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8dvms" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.320411 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.352796 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8lgxc"] Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.394361 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qg7wp"] Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.400827 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.402052 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qg7wp"] Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.471760 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.471872 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.471981 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-scripts\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.472092 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/274fba29-717c-46e8-89be-543c4540b871-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.472143 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.472275 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd9m5\" (UniqueName: \"kubernetes.io/projected/274fba29-717c-46e8-89be-543c4540b871-kube-api-access-rd9m5\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.518041 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.519472 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.522964 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.540233 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.577739 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.577793 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-scripts\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.577889 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.577961 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-config\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.577985 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/274fba29-717c-46e8-89be-543c4540b871-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.578051 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.578251 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.578279 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd9m5\" (UniqueName: \"kubernetes.io/projected/274fba29-717c-46e8-89be-543c4540b871-kube-api-access-rd9m5\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.578318 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.578340 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.578363 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.578379 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjmfw\" (UniqueName: \"kubernetes.io/projected/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-kube-api-access-bjmfw\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.578518 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/274fba29-717c-46e8-89be-543c4540b871-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.583425 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.585338 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.590482 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.595051 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-scripts\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.600505 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd9m5\" (UniqueName: \"kubernetes.io/projected/274fba29-717c-46e8-89be-543c4540b871-kube-api-access-rd9m5\") pod \"cinder-scheduler-0\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: E1210 09:52:44.680970 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8029fef4_8c21_42fd_8c79_7d76f6919d6e.slice/crio-4ed8aae26c2c29976b4530bb9e3d6e4715c5540045f2b3d25d7127bc18388151.scope\": RecentStats: unable to find data in memory cache]" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.681895 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-scripts\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682055 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-config\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682262 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data-custom\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682337 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4144d1bc-a1fa-4eea-904f-b87c770cec53-logs\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682395 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682420 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682463 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvd2k\" (UniqueName: \"kubernetes.io/projected/4144d1bc-a1fa-4eea-904f-b87c770cec53-kube-api-access-cvd2k\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682562 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682621 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4144d1bc-a1fa-4eea-904f-b87c770cec53-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682685 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682713 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.682750 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjmfw\" (UniqueName: \"kubernetes.io/projected/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-kube-api-access-bjmfw\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.683854 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.684247 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.684354 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-config\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.684367 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.685276 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.700058 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjmfw\" (UniqueName: \"kubernetes.io/projected/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-kube-api-access-bjmfw\") pod \"dnsmasq-dns-5c9776ccc5-qg7wp\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.785783 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-scripts\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.785929 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data-custom\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.785979 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4144d1bc-a1fa-4eea-904f-b87c770cec53-logs\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.786011 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.786037 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.786063 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvd2k\" (UniqueName: \"kubernetes.io/projected/4144d1bc-a1fa-4eea-904f-b87c770cec53-kube-api-access-cvd2k\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.786137 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4144d1bc-a1fa-4eea-904f-b87c770cec53-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.786302 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4144d1bc-a1fa-4eea-904f-b87c770cec53-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.791106 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4144d1bc-a1fa-4eea-904f-b87c770cec53-logs\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.793688 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.798471 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data-custom\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.798704 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.799656 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.799990 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-scripts\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.811640 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvd2k\" (UniqueName: \"kubernetes.io/projected/4144d1bc-a1fa-4eea-904f-b87c770cec53-kube-api-access-cvd2k\") pod \"cinder-api-0\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " pod="openstack/cinder-api-0" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.870384 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:44 crc kubenswrapper[4715]: I1210 09:52:44.880392 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.059933 4715 generic.go:334] "Generic (PLEG): container finished" podID="8029fef4-8c21-42fd-8c79-7d76f6919d6e" containerID="4ed8aae26c2c29976b4530bb9e3d6e4715c5540045f2b3d25d7127bc18388151" exitCode=0 Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.060260 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" event={"ID":"8029fef4-8c21-42fd-8c79-7d76f6919d6e","Type":"ContainerDied","Data":"4ed8aae26c2c29976b4530bb9e3d6e4715c5540045f2b3d25d7127bc18388151"} Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.060286 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" event={"ID":"8029fef4-8c21-42fd-8c79-7d76f6919d6e","Type":"ContainerStarted","Data":"da5fbc8731ccb67ab8304b58a3d1ec9c2f217cf010085ab7d6a5eb6d937413ab"} Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.067049 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86fc9d-jwnlv" event={"ID":"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2","Type":"ContainerStarted","Data":"44a195dfa862876a9a7e529edabaed0a7ac5415ff5d6db8d8ce8a825b5c92699"} Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.067088 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86fc9d-jwnlv" event={"ID":"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2","Type":"ContainerStarted","Data":"574f876f32cc798d7b9d401bb1023b5c1f7d1505ec08b920d6066d7071da7712"} Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.067097 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86fc9d-jwnlv" event={"ID":"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2","Type":"ContainerStarted","Data":"306cfa120f83e0c8818bde62ea10d26d1cdf4e7b3c1ce0ea8fa82bdbcefc152d"} Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.068666 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.068696 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.114303 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c86fc9d-jwnlv" podStartSLOduration=3.114283069 podStartE2EDuration="3.114283069s" podCreationTimestamp="2025-12-10 09:52:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:45.113377203 +0000 UTC m=+1127.856923464" watchObservedRunningTime="2025-12-10 09:52:45.114283069 +0000 UTC m=+1127.857829330" Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.384379 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.537865 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qg7wp"] Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.670347 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 09:52:45 crc kubenswrapper[4715]: W1210 09:52:45.766398 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4144d1bc_a1fa_4eea_904f_b87c770cec53.slice/crio-eeb7dc0cb0c53e5c30cad07fbf3316a9478cd5033c1b156c89583b0ac2b26042 WatchSource:0}: Error finding container eeb7dc0cb0c53e5c30cad07fbf3316a9478cd5033c1b156c89583b0ac2b26042: Status 404 returned error can't find the container with id eeb7dc0cb0c53e5c30cad07fbf3316a9478cd5033c1b156c89583b0ac2b26042 Dec 10 09:52:45 crc kubenswrapper[4715]: I1210 09:52:45.879395 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.031392 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-sb\") pod \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.031509 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-nb\") pod \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.031538 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppxmz\" (UniqueName: \"kubernetes.io/projected/8029fef4-8c21-42fd-8c79-7d76f6919d6e-kube-api-access-ppxmz\") pod \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.031629 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-svc\") pod \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.031725 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-swift-storage-0\") pod \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.031811 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-config\") pod \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\" (UID: \"8029fef4-8c21-42fd-8c79-7d76f6919d6e\") " Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.035430 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8029fef4-8c21-42fd-8c79-7d76f6919d6e-kube-api-access-ppxmz" (OuterVolumeSpecName: "kube-api-access-ppxmz") pod "8029fef4-8c21-42fd-8c79-7d76f6919d6e" (UID: "8029fef4-8c21-42fd-8c79-7d76f6919d6e"). InnerVolumeSpecName "kube-api-access-ppxmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.056272 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-config" (OuterVolumeSpecName: "config") pod "8029fef4-8c21-42fd-8c79-7d76f6919d6e" (UID: "8029fef4-8c21-42fd-8c79-7d76f6919d6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.057444 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8029fef4-8c21-42fd-8c79-7d76f6919d6e" (UID: "8029fef4-8c21-42fd-8c79-7d76f6919d6e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.060062 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8029fef4-8c21-42fd-8c79-7d76f6919d6e" (UID: "8029fef4-8c21-42fd-8c79-7d76f6919d6e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.063389 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8029fef4-8c21-42fd-8c79-7d76f6919d6e" (UID: "8029fef4-8c21-42fd-8c79-7d76f6919d6e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.068787 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8029fef4-8c21-42fd-8c79-7d76f6919d6e" (UID: "8029fef4-8c21-42fd-8c79-7d76f6919d6e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.079719 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" event={"ID":"8029fef4-8c21-42fd-8c79-7d76f6919d6e","Type":"ContainerDied","Data":"da5fbc8731ccb67ab8304b58a3d1ec9c2f217cf010085ab7d6a5eb6d937413ab"} Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.079763 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-8lgxc" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.079770 4715 scope.go:117] "RemoveContainer" containerID="4ed8aae26c2c29976b4530bb9e3d6e4715c5540045f2b3d25d7127bc18388151" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.088522 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4144d1bc-a1fa-4eea-904f-b87c770cec53","Type":"ContainerStarted","Data":"eeb7dc0cb0c53e5c30cad07fbf3316a9478cd5033c1b156c89583b0ac2b26042"} Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.091722 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" event={"ID":"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25","Type":"ContainerStarted","Data":"7dde7ea032f37ccc52e3c430b5de354a3b33123fbcf22765e9bed40c28d72cc8"} Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.093454 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"274fba29-717c-46e8-89be-543c4540b871","Type":"ContainerStarted","Data":"7d287713862915f0b996be130b4f7a0d283d8c0011a89d5e7c8a55e5668171df"} Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.134005 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.134034 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppxmz\" (UniqueName: \"kubernetes.io/projected/8029fef4-8c21-42fd-8c79-7d76f6919d6e-kube-api-access-ppxmz\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.134045 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.134055 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.134063 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.134071 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8029fef4-8c21-42fd-8c79-7d76f6919d6e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.167584 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8lgxc"] Dec 10 09:52:46 crc kubenswrapper[4715]: I1210 09:52:46.177461 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-8lgxc"] Dec 10 09:52:47 crc kubenswrapper[4715]: I1210 09:52:47.657903 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8029fef4-8c21-42fd-8c79-7d76f6919d6e" path="/var/lib/kubelet/pods/8029fef4-8c21-42fd-8c79-7d76f6919d6e/volumes" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.154288 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c964fbd47-67rln" event={"ID":"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b","Type":"ContainerStarted","Data":"382e1edb9a0021433b5952c50be5c303867633953c412fcbfa82741d29d4a22a"} Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.154664 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c964fbd47-67rln" event={"ID":"b46f5949-5bfa-4cfd-b70d-4d225ea12d6b","Type":"ContainerStarted","Data":"9247b30306937338c9bcee31721b3cdfe97fc3331b91aca20d759a51e4d99cdd"} Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.169743 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2167d98b-0365-421e-9d3d-6ff62f7bfb61","Type":"ContainerDied","Data":"d6a23411f595071e43d6ed0df1b67d55411b62de11ff04597a980928622d1e76"} Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.169835 4715 generic.go:334] "Generic (PLEG): container finished" podID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerID="d6a23411f595071e43d6ed0df1b67d55411b62de11ff04597a980928622d1e76" exitCode=0 Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.182285 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.186107 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5c964fbd47-67rln" podStartSLOduration=3.795848703 podStartE2EDuration="6.186078033s" podCreationTimestamp="2025-12-10 09:52:42 +0000 UTC" firstStartedPulling="2025-12-10 09:52:43.814371593 +0000 UTC m=+1126.557917844" lastFinishedPulling="2025-12-10 09:52:46.204600923 +0000 UTC m=+1128.948147174" observedRunningTime="2025-12-10 09:52:48.183585043 +0000 UTC m=+1130.927131294" watchObservedRunningTime="2025-12-10 09:52:48.186078033 +0000 UTC m=+1130.929624284" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.192799 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4144d1bc-a1fa-4eea-904f-b87c770cec53","Type":"ContainerStarted","Data":"5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8"} Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.192898 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.213296 4715 generic.go:334] "Generic (PLEG): container finished" podID="af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" containerID="4cf3d5f67480fb289d7598b0d540e92eaf74f47cb91e68dc19e767b6926f4794" exitCode=0 Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.213395 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" event={"ID":"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25","Type":"ContainerDied","Data":"4cf3d5f67480fb289d7598b0d540e92eaf74f47cb91e68dc19e767b6926f4794"} Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.223592 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" event={"ID":"73357509-8bf3-408e-9c35-ee6267593be1","Type":"ContainerStarted","Data":"2262e08232084955214bf0524b90e1a1eb0ee9cd08db19867c278e778fdb69a0"} Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.223630 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" event={"ID":"73357509-8bf3-408e-9c35-ee6267593be1","Type":"ContainerStarted","Data":"b0ad7020f55bc3de3dda3ba08207a07289c1eb244e33430917f441ad1fb48c95"} Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.228222 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.267341 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.267318683 podStartE2EDuration="4.267318683s" podCreationTimestamp="2025-12-10 09:52:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:48.254287714 +0000 UTC m=+1130.997833965" watchObservedRunningTime="2025-12-10 09:52:48.267318683 +0000 UTC m=+1131.010864934" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.284974 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-run-httpd\") pod \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.285072 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-combined-ca-bundle\") pod \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.285106 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czlpf\" (UniqueName: \"kubernetes.io/projected/2167d98b-0365-421e-9d3d-6ff62f7bfb61-kube-api-access-czlpf\") pod \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.285228 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-sg-core-conf-yaml\") pod \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.285303 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-config-data\") pod \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.285355 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-scripts\") pod \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.285430 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-log-httpd\") pod \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\" (UID: \"2167d98b-0365-421e-9d3d-6ff62f7bfb61\") " Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.290671 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2167d98b-0365-421e-9d3d-6ff62f7bfb61" (UID: "2167d98b-0365-421e-9d3d-6ff62f7bfb61"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.293116 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2167d98b-0365-421e-9d3d-6ff62f7bfb61" (UID: "2167d98b-0365-421e-9d3d-6ff62f7bfb61"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.322568 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-scripts" (OuterVolumeSpecName: "scripts") pod "2167d98b-0365-421e-9d3d-6ff62f7bfb61" (UID: "2167d98b-0365-421e-9d3d-6ff62f7bfb61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.323147 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6d588649f8-m9jx2" podStartSLOduration=3.829171557 podStartE2EDuration="6.323124883s" podCreationTimestamp="2025-12-10 09:52:42 +0000 UTC" firstStartedPulling="2025-12-10 09:52:43.713157698 +0000 UTC m=+1126.456703939" lastFinishedPulling="2025-12-10 09:52:46.207111014 +0000 UTC m=+1128.950657265" observedRunningTime="2025-12-10 09:52:48.315571879 +0000 UTC m=+1131.059118140" watchObservedRunningTime="2025-12-10 09:52:48.323124883 +0000 UTC m=+1131.066671134" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.366706 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2167d98b-0365-421e-9d3d-6ff62f7bfb61-kube-api-access-czlpf" (OuterVolumeSpecName: "kube-api-access-czlpf") pod "2167d98b-0365-421e-9d3d-6ff62f7bfb61" (UID: "2167d98b-0365-421e-9d3d-6ff62f7bfb61"). InnerVolumeSpecName "kube-api-access-czlpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.370161 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2167d98b-0365-421e-9d3d-6ff62f7bfb61" (UID: "2167d98b-0365-421e-9d3d-6ff62f7bfb61"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.391738 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.391778 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2167d98b-0365-421e-9d3d-6ff62f7bfb61-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.391788 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czlpf\" (UniqueName: \"kubernetes.io/projected/2167d98b-0365-421e-9d3d-6ff62f7bfb61-kube-api-access-czlpf\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.391800 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.391809 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.514591 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-config-data" (OuterVolumeSpecName: "config-data") pod "2167d98b-0365-421e-9d3d-6ff62f7bfb61" (UID: "2167d98b-0365-421e-9d3d-6ff62f7bfb61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.529125 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2167d98b-0365-421e-9d3d-6ff62f7bfb61" (UID: "2167d98b-0365-421e-9d3d-6ff62f7bfb61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.595514 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:48 crc kubenswrapper[4715]: I1210 09:52:48.595545 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2167d98b-0365-421e-9d3d-6ff62f7bfb61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.022336 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.265689 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"274fba29-717c-46e8-89be-543c4540b871","Type":"ContainerStarted","Data":"e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb"} Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.265733 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"274fba29-717c-46e8-89be-543c4540b871","Type":"ContainerStarted","Data":"d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3"} Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.268625 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.268617 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2167d98b-0365-421e-9d3d-6ff62f7bfb61","Type":"ContainerDied","Data":"d48c3f8c7911397fdca54dde3b2bac53d0fb6a4cd755e50728d3519f58e22c52"} Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.268752 4715 scope.go:117] "RemoveContainer" containerID="5bf233fe7cd9fee035d356ea9cfed7ecbd952a02f2fe9e5efc4c25f25096bb7d" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.272672 4715 generic.go:334] "Generic (PLEG): container finished" podID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerID="392365e94829e5068e3d8ff96fc73b2174fe79173b8c52c807194cfa86a86c43" exitCode=1 Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.272738 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4144d1bc-a1fa-4eea-904f-b87c770cec53","Type":"ContainerDied","Data":"392365e94829e5068e3d8ff96fc73b2174fe79173b8c52c807194cfa86a86c43"} Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.273383 4715 scope.go:117] "RemoveContainer" containerID="392365e94829e5068e3d8ff96fc73b2174fe79173b8c52c807194cfa86a86c43" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.293261 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" event={"ID":"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25","Type":"ContainerStarted","Data":"b0f842f4ca7b98a1bb87100e3351267e5cd609805b83ea442e3fad1d45914b7a"} Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.293298 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.327300 4715 scope.go:117] "RemoveContainer" containerID="d6790cb6e37145805338959fadfadc2c40fef0faaee91d9db225c0f632e72da0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.381048 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.763430688 podStartE2EDuration="5.381023068s" podCreationTimestamp="2025-12-10 09:52:44 +0000 UTC" firstStartedPulling="2025-12-10 09:52:45.756816477 +0000 UTC m=+1128.500362718" lastFinishedPulling="2025-12-10 09:52:47.374408847 +0000 UTC m=+1130.117955098" observedRunningTime="2025-12-10 09:52:49.327356139 +0000 UTC m=+1132.070902390" watchObservedRunningTime="2025-12-10 09:52:49.381023068 +0000 UTC m=+1132.124569319" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.403106 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" podStartSLOduration=5.403082472 podStartE2EDuration="5.403082472s" podCreationTimestamp="2025-12-10 09:52:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:49.359151359 +0000 UTC m=+1132.102697610" watchObservedRunningTime="2025-12-10 09:52:49.403082472 +0000 UTC m=+1132.146628733" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.430984 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.477117 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.494598 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:52:49 crc kubenswrapper[4715]: E1210 09:52:49.495120 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="ceilometer-notification-agent" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.495145 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="ceilometer-notification-agent" Dec 10 09:52:49 crc kubenswrapper[4715]: E1210 09:52:49.495182 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="sg-core" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.495191 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="sg-core" Dec 10 09:52:49 crc kubenswrapper[4715]: E1210 09:52:49.495207 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="proxy-httpd" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.495217 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="proxy-httpd" Dec 10 09:52:49 crc kubenswrapper[4715]: E1210 09:52:49.495229 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="ceilometer-central-agent" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.495237 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="ceilometer-central-agent" Dec 10 09:52:49 crc kubenswrapper[4715]: E1210 09:52:49.495252 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8029fef4-8c21-42fd-8c79-7d76f6919d6e" containerName="init" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.495259 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8029fef4-8c21-42fd-8c79-7d76f6919d6e" containerName="init" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.495492 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8029fef4-8c21-42fd-8c79-7d76f6919d6e" containerName="init" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.495533 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="proxy-httpd" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.495545 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="sg-core" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.495565 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="ceilometer-notification-agent" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.495577 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" containerName="ceilometer-central-agent" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.498982 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.502458 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.509342 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.509629 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.514856 4715 scope.go:117] "RemoveContainer" containerID="d6a23411f595071e43d6ed0df1b67d55411b62de11ff04597a980928622d1e76" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.584208 4715 scope.go:117] "RemoveContainer" containerID="382c1037c3c28ee70ddc99fb57d441e603a2485e73eef3c7410495d5fb30133b" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.592567 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.622856 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2167d98b-0365-421e-9d3d-6ff62f7bfb61" path="/var/lib/kubelet/pods/2167d98b-0365-421e-9d3d-6ff62f7bfb61/volumes" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.639868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-log-httpd\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.639949 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.639983 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2ggl\" (UniqueName: \"kubernetes.io/projected/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-kube-api-access-h2ggl\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.640022 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-config-data\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.640042 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-run-httpd\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.640094 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-scripts\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.640118 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.742465 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-scripts\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.742548 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.742695 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-log-httpd\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.742746 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.742778 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2ggl\" (UniqueName: \"kubernetes.io/projected/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-kube-api-access-h2ggl\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.742868 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-config-data\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.742967 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-run-httpd\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.745470 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-run-httpd\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.746125 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-log-httpd\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.749151 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-config-data\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.749821 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.749886 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.777779 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-scripts\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.784852 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2ggl\" (UniqueName: \"kubernetes.io/projected/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-kube-api-access-h2ggl\") pod \"ceilometer-0\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.794203 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.870488 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.927634 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.983988 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7856564b58-9kr94"] Dec 10 09:52:49 crc kubenswrapper[4715]: I1210 09:52:49.998216 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.001097 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.007280 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.052983 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7856564b58-9kr94"] Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.178451 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-config-data\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.178659 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56h49\" (UniqueName: \"kubernetes.io/projected/07a44832-3242-42dd-9366-5628181a6c6e-kube-api-access-56h49\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.178730 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07a44832-3242-42dd-9366-5628181a6c6e-logs\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.178789 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-config-data-custom\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.178812 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-combined-ca-bundle\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.178857 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-internal-tls-certs\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.178872 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-public-tls-certs\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.285959 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-config-data\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.286000 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56h49\" (UniqueName: \"kubernetes.io/projected/07a44832-3242-42dd-9366-5628181a6c6e-kube-api-access-56h49\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.286071 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07a44832-3242-42dd-9366-5628181a6c6e-logs\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.286130 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-config-data-custom\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.286153 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-combined-ca-bundle\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.286192 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-internal-tls-certs\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.286209 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-public-tls-certs\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.286855 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07a44832-3242-42dd-9366-5628181a6c6e-logs\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.296380 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-public-tls-certs\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.298466 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-internal-tls-certs\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.299443 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-combined-ca-bundle\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.311602 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-config-data\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.318141 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56h49\" (UniqueName: \"kubernetes.io/projected/07a44832-3242-42dd-9366-5628181a6c6e-kube-api-access-56h49\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.329201 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07a44832-3242-42dd-9366-5628181a6c6e-config-data-custom\") pod \"barbican-api-7856564b58-9kr94\" (UID: \"07a44832-3242-42dd-9366-5628181a6c6e\") " pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.490123 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:52:50 crc kubenswrapper[4715]: W1210 09:52:50.490342 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbacefd07_78b8_45cb_930b_5c4ec9fa9b95.slice/crio-01eada311fa18bba2b676fad1667ca814b8f7af89bf8f8762f840e14863420ca WatchSource:0}: Error finding container 01eada311fa18bba2b676fad1667ca814b8f7af89bf8f8762f840e14863420ca: Status 404 returned error can't find the container with id 01eada311fa18bba2b676fad1667ca814b8f7af89bf8f8762f840e14863420ca Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.628221 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:50 crc kubenswrapper[4715]: I1210 09:52:50.881359 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-api-0" Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.157519 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7856564b58-9kr94"] Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.333740 4715 generic.go:334] "Generic (PLEG): container finished" podID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerID="79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa" exitCode=1 Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.333809 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4144d1bc-a1fa-4eea-904f-b87c770cec53","Type":"ContainerDied","Data":"79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa"} Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.333847 4715 scope.go:117] "RemoveContainer" containerID="392365e94829e5068e3d8ff96fc73b2174fe79173b8c52c807194cfa86a86c43" Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.334063 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerName="cinder-api-log" containerID="cri-o://5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8" gracePeriod=30 Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.343156 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7856564b58-9kr94" event={"ID":"07a44832-3242-42dd-9366-5628181a6c6e","Type":"ContainerStarted","Data":"453ed5b81a4cd934b02c08e3d6cbbd4a9bfa9644ddc17c6a6edd6198f386fa7a"} Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.351130 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bacefd07-78b8-45cb-930b-5c4ec9fa9b95","Type":"ContainerStarted","Data":"01eada311fa18bba2b676fad1667ca814b8f7af89bf8f8762f840e14863420ca"} Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.707407 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-85df69bf85-6cwpq" Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.781085 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6c77448d4b-7p5j8"] Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.784809 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6c77448d4b-7p5j8" podUID="96bdc06f-533b-4ba3-8074-a92466586d44" containerName="neutron-api" containerID="cri-o://06577babfec20c9050142133b19e2ff06bc98a9bdd42bed1ac09c8b05f0f3ed8" gracePeriod=30 Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.785104 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6c77448d4b-7p5j8" podUID="96bdc06f-533b-4ba3-8074-a92466586d44" containerName="neutron-httpd" containerID="cri-o://f3063c5b676bdfc3fdb327164846de92265a2df2487ea64ff436fc663bdc6360" gracePeriod=30 Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.897510 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-857b4dfcd4-wnlbz" Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.919928 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:52:51 crc kubenswrapper[4715]: I1210 09:52:51.957878 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67f6cc7984-tshct"] Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.220372 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.326499 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data\") pod \"4144d1bc-a1fa-4eea-904f-b87c770cec53\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.326569 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data-custom\") pod \"4144d1bc-a1fa-4eea-904f-b87c770cec53\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.326628 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4144d1bc-a1fa-4eea-904f-b87c770cec53-etc-machine-id\") pod \"4144d1bc-a1fa-4eea-904f-b87c770cec53\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.326658 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvd2k\" (UniqueName: \"kubernetes.io/projected/4144d1bc-a1fa-4eea-904f-b87c770cec53-kube-api-access-cvd2k\") pod \"4144d1bc-a1fa-4eea-904f-b87c770cec53\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.326688 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-combined-ca-bundle\") pod \"4144d1bc-a1fa-4eea-904f-b87c770cec53\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.326723 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-scripts\") pod \"4144d1bc-a1fa-4eea-904f-b87c770cec53\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.326747 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4144d1bc-a1fa-4eea-904f-b87c770cec53-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4144d1bc-a1fa-4eea-904f-b87c770cec53" (UID: "4144d1bc-a1fa-4eea-904f-b87c770cec53"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.326795 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4144d1bc-a1fa-4eea-904f-b87c770cec53-logs\") pod \"4144d1bc-a1fa-4eea-904f-b87c770cec53\" (UID: \"4144d1bc-a1fa-4eea-904f-b87c770cec53\") " Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.327344 4715 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4144d1bc-a1fa-4eea-904f-b87c770cec53-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.327556 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4144d1bc-a1fa-4eea-904f-b87c770cec53-logs" (OuterVolumeSpecName: "logs") pod "4144d1bc-a1fa-4eea-904f-b87c770cec53" (UID: "4144d1bc-a1fa-4eea-904f-b87c770cec53"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.336245 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4144d1bc-a1fa-4eea-904f-b87c770cec53-kube-api-access-cvd2k" (OuterVolumeSpecName: "kube-api-access-cvd2k") pod "4144d1bc-a1fa-4eea-904f-b87c770cec53" (UID: "4144d1bc-a1fa-4eea-904f-b87c770cec53"). InnerVolumeSpecName "kube-api-access-cvd2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.341334 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4144d1bc-a1fa-4eea-904f-b87c770cec53" (UID: "4144d1bc-a1fa-4eea-904f-b87c770cec53"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.345057 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-scripts" (OuterVolumeSpecName: "scripts") pod "4144d1bc-a1fa-4eea-904f-b87c770cec53" (UID: "4144d1bc-a1fa-4eea-904f-b87c770cec53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.377121 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4144d1bc-a1fa-4eea-904f-b87c770cec53" (UID: "4144d1bc-a1fa-4eea-904f-b87c770cec53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.381201 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7856564b58-9kr94" event={"ID":"07a44832-3242-42dd-9366-5628181a6c6e","Type":"ContainerStarted","Data":"73150973ad93f42ae6131678c2e220ccdb1a40e4c8ccc0efa91b49a2bec50a0d"} Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.381251 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7856564b58-9kr94" event={"ID":"07a44832-3242-42dd-9366-5628181a6c6e","Type":"ContainerStarted","Data":"d3841a2ad5c680739eba1df4f6f82247b78ea475ae00c5bda1457996190ce803"} Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.382509 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.382537 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.397179 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bacefd07-78b8-45cb-930b-5c4ec9fa9b95","Type":"ContainerStarted","Data":"fdffd979d1ec3cb49ed1246f33350d6b6b6238dca028a1073765d92ce628d7cb"} Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.410083 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7856564b58-9kr94" podStartSLOduration=3.410062151 podStartE2EDuration="3.410062151s" podCreationTimestamp="2025-12-10 09:52:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:52.406236723 +0000 UTC m=+1135.149782984" watchObservedRunningTime="2025-12-10 09:52:52.410062151 +0000 UTC m=+1135.153608402" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.411213 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data" (OuterVolumeSpecName: "config-data") pod "4144d1bc-a1fa-4eea-904f-b87c770cec53" (UID: "4144d1bc-a1fa-4eea-904f-b87c770cec53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.412210 4715 generic.go:334] "Generic (PLEG): container finished" podID="96bdc06f-533b-4ba3-8074-a92466586d44" containerID="f3063c5b676bdfc3fdb327164846de92265a2df2487ea64ff436fc663bdc6360" exitCode=0 Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.412342 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c77448d4b-7p5j8" event={"ID":"96bdc06f-533b-4ba3-8074-a92466586d44","Type":"ContainerDied","Data":"f3063c5b676bdfc3fdb327164846de92265a2df2487ea64ff436fc663bdc6360"} Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.418557 4715 generic.go:334] "Generic (PLEG): container finished" podID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerID="5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8" exitCode=143 Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.418694 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.418686 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4144d1bc-a1fa-4eea-904f-b87c770cec53","Type":"ContainerDied","Data":"5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8"} Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.418752 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4144d1bc-a1fa-4eea-904f-b87c770cec53","Type":"ContainerDied","Data":"eeb7dc0cb0c53e5c30cad07fbf3316a9478cd5033c1b156c89583b0ac2b26042"} Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.418775 4715 scope.go:117] "RemoveContainer" containerID="79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.418831 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67f6cc7984-tshct" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon-log" containerID="cri-o://c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171" gracePeriod=30 Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.419219 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67f6cc7984-tshct" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon" containerID="cri-o://e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d" gracePeriod=30 Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.429221 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvd2k\" (UniqueName: \"kubernetes.io/projected/4144d1bc-a1fa-4eea-904f-b87c770cec53-kube-api-access-cvd2k\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.429247 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.429259 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.429268 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4144d1bc-a1fa-4eea-904f-b87c770cec53-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.429276 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.429285 4715 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4144d1bc-a1fa-4eea-904f-b87c770cec53-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.534690 4715 scope.go:117] "RemoveContainer" containerID="5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.595728 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.610673 4715 scope.go:117] "RemoveContainer" containerID="79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa" Dec 10 09:52:52 crc kubenswrapper[4715]: E1210 09:52:52.614665 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa\": container with ID starting with 79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa not found: ID does not exist" containerID="79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.614772 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa"} err="failed to get container status \"79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa\": rpc error: code = NotFound desc = could not find container \"79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa\": container with ID starting with 79f9e3543583e0a1729e6fcd56bbb2f6470c3902c8d9aab7b2f81bca8ffc2aaa not found: ID does not exist" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.614817 4715 scope.go:117] "RemoveContainer" containerID="5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8" Dec 10 09:52:52 crc kubenswrapper[4715]: E1210 09:52:52.616248 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8\": container with ID starting with 5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8 not found: ID does not exist" containerID="5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.616272 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8"} err="failed to get container status \"5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8\": rpc error: code = NotFound desc = could not find container \"5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8\": container with ID starting with 5a4256915d95b96e2dbcc783d86ac3486223d08b8a25edab58cae36f7580ddd8 not found: ID does not exist" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.638812 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.670766 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 10 09:52:52 crc kubenswrapper[4715]: E1210 09:52:52.671562 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerName="cinder-api" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.671589 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerName="cinder-api" Dec 10 09:52:52 crc kubenswrapper[4715]: E1210 09:52:52.671610 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerName="cinder-api" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.671618 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerName="cinder-api" Dec 10 09:52:52 crc kubenswrapper[4715]: E1210 09:52:52.671633 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerName="cinder-api-log" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.671641 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerName="cinder-api-log" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.671899 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerName="cinder-api-log" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.671952 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerName="cinder-api" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.671969 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" containerName="cinder-api" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.673320 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.679265 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.679516 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.679623 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.690495 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.736281 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.737087 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzxjn\" (UniqueName: \"kubernetes.io/projected/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-kube-api-access-qzxjn\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.737248 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.737463 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.737576 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-scripts\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.738124 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.738238 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-config-data-custom\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.738458 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-logs\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.738650 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-config-data\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.841796 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.842237 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzxjn\" (UniqueName: \"kubernetes.io/projected/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-kube-api-access-qzxjn\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.842353 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.842560 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.843357 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-scripts\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.843595 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.843699 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-config-data-custom\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.843847 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-logs\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.844006 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.844148 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-config-data\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.844636 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-logs\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.851929 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.852400 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.852811 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-config-data\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.856434 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-scripts\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.860491 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-config-data-custom\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.871451 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:52 crc kubenswrapper[4715]: I1210 09:52:52.876408 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzxjn\" (UniqueName: \"kubernetes.io/projected/e911669d-5a8e-49d6-a6e2-b66dfae8f7e5-kube-api-access-qzxjn\") pod \"cinder-api-0\" (UID: \"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5\") " pod="openstack/cinder-api-0" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.005818 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.436829 4715 generic.go:334] "Generic (PLEG): container finished" podID="96bdc06f-533b-4ba3-8074-a92466586d44" containerID="06577babfec20c9050142133b19e2ff06bc98a9bdd42bed1ac09c8b05f0f3ed8" exitCode=0 Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.436902 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c77448d4b-7p5j8" event={"ID":"96bdc06f-533b-4ba3-8074-a92466586d44","Type":"ContainerDied","Data":"06577babfec20c9050142133b19e2ff06bc98a9bdd42bed1ac09c8b05f0f3ed8"} Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.447008 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bacefd07-78b8-45cb-930b-5c4ec9fa9b95","Type":"ContainerStarted","Data":"67ce57ae17678110c220b601a055667d6f51c6269cad420b147e48a7fbf99527"} Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.609060 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.621265 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4144d1bc-a1fa-4eea-904f-b87c770cec53" path="/var/lib/kubelet/pods/4144d1bc-a1fa-4eea-904f-b87c770cec53/volumes" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.713436 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.787112 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-httpd-config\") pod \"96bdc06f-533b-4ba3-8074-a92466586d44\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.787171 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfjvb\" (UniqueName: \"kubernetes.io/projected/96bdc06f-533b-4ba3-8074-a92466586d44-kube-api-access-vfjvb\") pod \"96bdc06f-533b-4ba3-8074-a92466586d44\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.787248 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-config\") pod \"96bdc06f-533b-4ba3-8074-a92466586d44\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.787285 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-ovndb-tls-certs\") pod \"96bdc06f-533b-4ba3-8074-a92466586d44\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.787327 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-combined-ca-bundle\") pod \"96bdc06f-533b-4ba3-8074-a92466586d44\" (UID: \"96bdc06f-533b-4ba3-8074-a92466586d44\") " Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.794489 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "96bdc06f-533b-4ba3-8074-a92466586d44" (UID: "96bdc06f-533b-4ba3-8074-a92466586d44"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.798161 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96bdc06f-533b-4ba3-8074-a92466586d44-kube-api-access-vfjvb" (OuterVolumeSpecName: "kube-api-access-vfjvb") pod "96bdc06f-533b-4ba3-8074-a92466586d44" (UID: "96bdc06f-533b-4ba3-8074-a92466586d44"). InnerVolumeSpecName "kube-api-access-vfjvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.891969 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.892025 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfjvb\" (UniqueName: \"kubernetes.io/projected/96bdc06f-533b-4ba3-8074-a92466586d44-kube-api-access-vfjvb\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.898875 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96bdc06f-533b-4ba3-8074-a92466586d44" (UID: "96bdc06f-533b-4ba3-8074-a92466586d44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.906273 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-config" (OuterVolumeSpecName: "config") pod "96bdc06f-533b-4ba3-8074-a92466586d44" (UID: "96bdc06f-533b-4ba3-8074-a92466586d44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.948006 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "96bdc06f-533b-4ba3-8074-a92466586d44" (UID: "96bdc06f-533b-4ba3-8074-a92466586d44"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.994940 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.994988 4715 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:53 crc kubenswrapper[4715]: I1210 09:52:53.995033 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96bdc06f-533b-4ba3-8074-a92466586d44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.458211 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5","Type":"ContainerStarted","Data":"bb0d4e3c68aa282d48ec842bb4ac2d41bbd8d4d7d94bf1911a2f51fd9e18184d"} Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.483670 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bacefd07-78b8-45cb-930b-5c4ec9fa9b95","Type":"ContainerStarted","Data":"bb5112104ae2dcd87b19d76c289b8d38c5c89a6ab6e2f75f72729cf6a5a8062a"} Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.493562 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c77448d4b-7p5j8" Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.493635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c77448d4b-7p5j8" event={"ID":"96bdc06f-533b-4ba3-8074-a92466586d44","Type":"ContainerDied","Data":"7dee8d82e7b9b5c6a3321400f48ca87b56e22c283c551b61c4a9fdbc01bc8e4f"} Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.493729 4715 scope.go:117] "RemoveContainer" containerID="f3063c5b676bdfc3fdb327164846de92265a2df2487ea64ff436fc663bdc6360" Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.532969 4715 scope.go:117] "RemoveContainer" containerID="06577babfec20c9050142133b19e2ff06bc98a9bdd42bed1ac09c8b05f0f3ed8" Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.544428 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6c77448d4b-7p5j8"] Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.551527 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6c77448d4b-7p5j8"] Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.872060 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.976759 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-t8855"] Dec 10 09:52:54 crc kubenswrapper[4715]: I1210 09:52:54.977197 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-t8855" podUID="296b014c-dc51-422d-ab1a-3e2018759e4a" containerName="dnsmasq-dns" containerID="cri-o://77d73f827aff7bd2ee4d514c4e9756d7255049a02f1d032a89c5b6b1b046e946" gracePeriod=10 Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.055548 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.186230 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.267275 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.380982 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.504311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5","Type":"ContainerStarted","Data":"68cb9bbe62b454a45650ef8dcfe49d8739c0072369df33802bbea893ce3500c1"} Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.504551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e911669d-5a8e-49d6-a6e2-b66dfae8f7e5","Type":"ContainerStarted","Data":"246694d3698d0b0a523a155b6268a128237e24babd78cb52fddcc1437fea850a"} Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.504587 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.511808 4715 generic.go:334] "Generic (PLEG): container finished" podID="296b014c-dc51-422d-ab1a-3e2018759e4a" containerID="77d73f827aff7bd2ee4d514c4e9756d7255049a02f1d032a89c5b6b1b046e946" exitCode=0 Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.512091 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-t8855" event={"ID":"296b014c-dc51-422d-ab1a-3e2018759e4a","Type":"ContainerDied","Data":"77d73f827aff7bd2ee4d514c4e9756d7255049a02f1d032a89c5b6b1b046e946"} Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.512131 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-t8855" event={"ID":"296b014c-dc51-422d-ab1a-3e2018759e4a","Type":"ContainerDied","Data":"84db2e84a14905e893ae687f3ce9ced9a715c8ef9230e64405add47dd1075e80"} Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.512144 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84db2e84a14905e893ae687f3ce9ced9a715c8ef9230e64405add47dd1075e80" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.512319 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="274fba29-717c-46e8-89be-543c4540b871" containerName="cinder-scheduler" containerID="cri-o://e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb" gracePeriod=30 Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.512440 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="274fba29-717c-46e8-89be-543c4540b871" containerName="probe" containerID="cri-o://d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3" gracePeriod=30 Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.542281 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.542258405 podStartE2EDuration="3.542258405s" podCreationTimestamp="2025-12-10 09:52:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:52:55.524603985 +0000 UTC m=+1138.268150296" watchObservedRunningTime="2025-12-10 09:52:55.542258405 +0000 UTC m=+1138.285804676" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.584357 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.631824 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96bdc06f-533b-4ba3-8074-a92466586d44" path="/var/lib/kubelet/pods/96bdc06f-533b-4ba3-8074-a92466586d44/volumes" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.745129 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-swift-storage-0\") pod \"296b014c-dc51-422d-ab1a-3e2018759e4a\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.745266 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jb76t\" (UniqueName: \"kubernetes.io/projected/296b014c-dc51-422d-ab1a-3e2018759e4a-kube-api-access-jb76t\") pod \"296b014c-dc51-422d-ab1a-3e2018759e4a\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.745370 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-nb\") pod \"296b014c-dc51-422d-ab1a-3e2018759e4a\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.745401 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-svc\") pod \"296b014c-dc51-422d-ab1a-3e2018759e4a\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.745532 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-config\") pod \"296b014c-dc51-422d-ab1a-3e2018759e4a\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.745569 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-sb\") pod \"296b014c-dc51-422d-ab1a-3e2018759e4a\" (UID: \"296b014c-dc51-422d-ab1a-3e2018759e4a\") " Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.752550 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/296b014c-dc51-422d-ab1a-3e2018759e4a-kube-api-access-jb76t" (OuterVolumeSpecName: "kube-api-access-jb76t") pod "296b014c-dc51-422d-ab1a-3e2018759e4a" (UID: "296b014c-dc51-422d-ab1a-3e2018759e4a"). InnerVolumeSpecName "kube-api-access-jb76t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.817053 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "296b014c-dc51-422d-ab1a-3e2018759e4a" (UID: "296b014c-dc51-422d-ab1a-3e2018759e4a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.849113 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.849145 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jb76t\" (UniqueName: \"kubernetes.io/projected/296b014c-dc51-422d-ab1a-3e2018759e4a-kube-api-access-jb76t\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.859569 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-config" (OuterVolumeSpecName: "config") pod "296b014c-dc51-422d-ab1a-3e2018759e4a" (UID: "296b014c-dc51-422d-ab1a-3e2018759e4a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.880041 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "296b014c-dc51-422d-ab1a-3e2018759e4a" (UID: "296b014c-dc51-422d-ab1a-3e2018759e4a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.883722 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "296b014c-dc51-422d-ab1a-3e2018759e4a" (UID: "296b014c-dc51-422d-ab1a-3e2018759e4a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.910899 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "296b014c-dc51-422d-ab1a-3e2018759e4a" (UID: "296b014c-dc51-422d-ab1a-3e2018759e4a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.951132 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.951178 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.951193 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:55 crc kubenswrapper[4715]: I1210 09:52:55.951204 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/296b014c-dc51-422d-ab1a-3e2018759e4a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.529570 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bacefd07-78b8-45cb-930b-5c4ec9fa9b95","Type":"ContainerStarted","Data":"bbdba2f1b665159bcef79064a0181735663c38eba05c61cea686e9ca34a9f6ac"} Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.532071 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.534603 4715 generic.go:334] "Generic (PLEG): container finished" podID="274fba29-717c-46e8-89be-543c4540b871" containerID="d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3" exitCode=0 Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.534649 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"274fba29-717c-46e8-89be-543c4540b871","Type":"ContainerDied","Data":"d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3"} Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.538527 4715 generic.go:334] "Generic (PLEG): container finished" podID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerID="e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d" exitCode=0 Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.538679 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-t8855" Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.538806 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f6cc7984-tshct" event={"ID":"5a1c5abd-cfc5-427f-a378-cef6d31deb0e","Type":"ContainerDied","Data":"e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d"} Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.578299 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.309643612 podStartE2EDuration="7.578275861s" podCreationTimestamp="2025-12-10 09:52:49 +0000 UTC" firstStartedPulling="2025-12-10 09:52:50.492296705 +0000 UTC m=+1133.235842956" lastFinishedPulling="2025-12-10 09:52:55.760928954 +0000 UTC m=+1138.504475205" observedRunningTime="2025-12-10 09:52:56.574110203 +0000 UTC m=+1139.317656474" watchObservedRunningTime="2025-12-10 09:52:56.578275861 +0000 UTC m=+1139.321822122" Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.693334 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-t8855"] Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.703438 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-t8855"] Dec 10 09:52:56 crc kubenswrapper[4715]: I1210 09:52:56.861204 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67f6cc7984-tshct" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.549126 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.550189 4715 generic.go:334] "Generic (PLEG): container finished" podID="274fba29-717c-46e8-89be-543c4540b871" containerID="e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb" exitCode=0 Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.550931 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"274fba29-717c-46e8-89be-543c4540b871","Type":"ContainerDied","Data":"e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb"} Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.550987 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"274fba29-717c-46e8-89be-543c4540b871","Type":"ContainerDied","Data":"7d287713862915f0b996be130b4f7a0d283d8c0011a89d5e7c8a55e5668171df"} Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.551009 4715 scope.go:117] "RemoveContainer" containerID="d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.615430 4715 scope.go:117] "RemoveContainer" containerID="e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.648360 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="296b014c-dc51-422d-ab1a-3e2018759e4a" path="/var/lib/kubelet/pods/296b014c-dc51-422d-ab1a-3e2018759e4a/volumes" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.676377 4715 scope.go:117] "RemoveContainer" containerID="d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3" Dec 10 09:52:57 crc kubenswrapper[4715]: E1210 09:52:57.679207 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3\": container with ID starting with d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3 not found: ID does not exist" containerID="d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.679562 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3"} err="failed to get container status \"d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3\": rpc error: code = NotFound desc = could not find container \"d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3\": container with ID starting with d5da699ccc44298dc2484892a1c748d3ee2c652f79465eee2f72c968c8e96ec3 not found: ID does not exist" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.679602 4715 scope.go:117] "RemoveContainer" containerID="e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb" Dec 10 09:52:57 crc kubenswrapper[4715]: E1210 09:52:57.685339 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb\": container with ID starting with e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb not found: ID does not exist" containerID="e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.685388 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb"} err="failed to get container status \"e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb\": rpc error: code = NotFound desc = could not find container \"e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb\": container with ID starting with e7af9cbb13535f2ab74a59249dc0f9e994b41eb28e56c949fc5f861154010adb not found: ID does not exist" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.702636 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-scripts\") pod \"274fba29-717c-46e8-89be-543c4540b871\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.702685 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data\") pod \"274fba29-717c-46e8-89be-543c4540b871\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.702873 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd9m5\" (UniqueName: \"kubernetes.io/projected/274fba29-717c-46e8-89be-543c4540b871-kube-api-access-rd9m5\") pod \"274fba29-717c-46e8-89be-543c4540b871\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.702899 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/274fba29-717c-46e8-89be-543c4540b871-etc-machine-id\") pod \"274fba29-717c-46e8-89be-543c4540b871\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.703000 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data-custom\") pod \"274fba29-717c-46e8-89be-543c4540b871\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.703053 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-combined-ca-bundle\") pod \"274fba29-717c-46e8-89be-543c4540b871\" (UID: \"274fba29-717c-46e8-89be-543c4540b871\") " Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.704034 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/274fba29-717c-46e8-89be-543c4540b871-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "274fba29-717c-46e8-89be-543c4540b871" (UID: "274fba29-717c-46e8-89be-543c4540b871"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.705235 4715 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/274fba29-717c-46e8-89be-543c4540b871-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.710092 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-scripts" (OuterVolumeSpecName: "scripts") pod "274fba29-717c-46e8-89be-543c4540b871" (UID: "274fba29-717c-46e8-89be-543c4540b871"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.711828 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/274fba29-717c-46e8-89be-543c4540b871-kube-api-access-rd9m5" (OuterVolumeSpecName: "kube-api-access-rd9m5") pod "274fba29-717c-46e8-89be-543c4540b871" (UID: "274fba29-717c-46e8-89be-543c4540b871"). InnerVolumeSpecName "kube-api-access-rd9m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.721091 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "274fba29-717c-46e8-89be-543c4540b871" (UID: "274fba29-717c-46e8-89be-543c4540b871"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.772549 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "274fba29-717c-46e8-89be-543c4540b871" (UID: "274fba29-717c-46e8-89be-543c4540b871"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.808827 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.808891 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd9m5\" (UniqueName: \"kubernetes.io/projected/274fba29-717c-46e8-89be-543c4540b871-kube-api-access-rd9m5\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.808950 4715 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.808964 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.836055 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data" (OuterVolumeSpecName: "config-data") pod "274fba29-717c-46e8-89be-543c4540b871" (UID: "274fba29-717c-46e8-89be-543c4540b871"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.910527 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/274fba29-717c-46e8-89be-543c4540b871-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:52:57 crc kubenswrapper[4715]: I1210 09:52:57.968412 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.168050 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5777d57b56-6lbcp" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.559479 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.613492 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.623679 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.647381 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 09:52:58 crc kubenswrapper[4715]: E1210 09:52:58.648004 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="296b014c-dc51-422d-ab1a-3e2018759e4a" containerName="init" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648024 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="296b014c-dc51-422d-ab1a-3e2018759e4a" containerName="init" Dec 10 09:52:58 crc kubenswrapper[4715]: E1210 09:52:58.648053 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274fba29-717c-46e8-89be-543c4540b871" containerName="cinder-scheduler" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648060 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="274fba29-717c-46e8-89be-543c4540b871" containerName="cinder-scheduler" Dec 10 09:52:58 crc kubenswrapper[4715]: E1210 09:52:58.648082 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96bdc06f-533b-4ba3-8074-a92466586d44" containerName="neutron-api" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648089 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="96bdc06f-533b-4ba3-8074-a92466586d44" containerName="neutron-api" Dec 10 09:52:58 crc kubenswrapper[4715]: E1210 09:52:58.648100 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="296b014c-dc51-422d-ab1a-3e2018759e4a" containerName="dnsmasq-dns" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648105 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="296b014c-dc51-422d-ab1a-3e2018759e4a" containerName="dnsmasq-dns" Dec 10 09:52:58 crc kubenswrapper[4715]: E1210 09:52:58.648119 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274fba29-717c-46e8-89be-543c4540b871" containerName="probe" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648124 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="274fba29-717c-46e8-89be-543c4540b871" containerName="probe" Dec 10 09:52:58 crc kubenswrapper[4715]: E1210 09:52:58.648133 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96bdc06f-533b-4ba3-8074-a92466586d44" containerName="neutron-httpd" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648139 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="96bdc06f-533b-4ba3-8074-a92466586d44" containerName="neutron-httpd" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648301 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="274fba29-717c-46e8-89be-543c4540b871" containerName="cinder-scheduler" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648317 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="96bdc06f-533b-4ba3-8074-a92466586d44" containerName="neutron-api" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648325 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="296b014c-dc51-422d-ab1a-3e2018759e4a" containerName="dnsmasq-dns" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648345 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="96bdc06f-533b-4ba3-8074-a92466586d44" containerName="neutron-httpd" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.648358 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="274fba29-717c-46e8-89be-543c4540b871" containerName="probe" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.649375 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.652484 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.670137 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.828668 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-scripts\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.828716 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.828749 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l9hx\" (UniqueName: \"kubernetes.io/projected/807957b7-0d34-49ab-8b79-6abcd56ad840-kube-api-access-2l9hx\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.828791 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-config-data\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.828809 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/807957b7-0d34-49ab-8b79-6abcd56ad840-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.828830 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.930785 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-scripts\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.930831 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.930868 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l9hx\" (UniqueName: \"kubernetes.io/projected/807957b7-0d34-49ab-8b79-6abcd56ad840-kube-api-access-2l9hx\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.930921 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-config-data\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.930940 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/807957b7-0d34-49ab-8b79-6abcd56ad840-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.930961 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.931904 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/807957b7-0d34-49ab-8b79-6abcd56ad840-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.936609 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-config-data\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.937028 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.943326 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-scripts\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.944377 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807957b7-0d34-49ab-8b79-6abcd56ad840-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.952710 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l9hx\" (UniqueName: \"kubernetes.io/projected/807957b7-0d34-49ab-8b79-6abcd56ad840-kube-api-access-2l9hx\") pod \"cinder-scheduler-0\" (UID: \"807957b7-0d34-49ab-8b79-6abcd56ad840\") " pod="openstack/cinder-scheduler-0" Dec 10 09:52:58 crc kubenswrapper[4715]: I1210 09:52:58.967192 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 09:52:59 crc kubenswrapper[4715]: W1210 09:52:59.422679 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod807957b7_0d34_49ab_8b79_6abcd56ad840.slice/crio-0120d7e75cd8374cb4b1305b62bef82e484e876a5395485d0520870913b21586 WatchSource:0}: Error finding container 0120d7e75cd8374cb4b1305b62bef82e484e876a5395485d0520870913b21586: Status 404 returned error can't find the container with id 0120d7e75cd8374cb4b1305b62bef82e484e876a5395485d0520870913b21586 Dec 10 09:52:59 crc kubenswrapper[4715]: I1210 09:52:59.426470 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 09:52:59 crc kubenswrapper[4715]: I1210 09:52:59.569749 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"807957b7-0d34-49ab-8b79-6abcd56ad840","Type":"ContainerStarted","Data":"0120d7e75cd8374cb4b1305b62bef82e484e876a5395485d0520870913b21586"} Dec 10 09:52:59 crc kubenswrapper[4715]: I1210 09:52:59.626070 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="274fba29-717c-46e8-89be-543c4540b871" path="/var/lib/kubelet/pods/274fba29-717c-46e8-89be-543c4540b871/volumes" Dec 10 09:53:00 crc kubenswrapper[4715]: I1210 09:53:00.585840 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"807957b7-0d34-49ab-8b79-6abcd56ad840","Type":"ContainerStarted","Data":"0f08c82d944867400ba9bbac162437b1f76ee9027c6932cd1dc725f1f2f4f0ed"} Dec 10 09:53:01 crc kubenswrapper[4715]: I1210 09:53:01.594852 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"807957b7-0d34-49ab-8b79-6abcd56ad840","Type":"ContainerStarted","Data":"64ba5fad1014f4c58e13d439f63edef1db69fd42556431a1e8397145ba4917e1"} Dec 10 09:53:02 crc kubenswrapper[4715]: I1210 09:53:02.160492 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:53:02 crc kubenswrapper[4715]: I1210 09:53:02.183732 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.183714035 podStartE2EDuration="4.183714035s" podCreationTimestamp="2025-12-10 09:52:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:53:01.627384807 +0000 UTC m=+1144.370931058" watchObservedRunningTime="2025-12-10 09:53:02.183714035 +0000 UTC m=+1144.927260286" Dec 10 09:53:02 crc kubenswrapper[4715]: I1210 09:53:02.204714 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7856564b58-9kr94" Dec 10 09:53:02 crc kubenswrapper[4715]: I1210 09:53:02.274897 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c86fc9d-jwnlv"] Dec 10 09:53:02 crc kubenswrapper[4715]: I1210 09:53:02.275190 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c86fc9d-jwnlv" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerName="barbican-api-log" containerID="cri-o://574f876f32cc798d7b9d401bb1023b5c1f7d1505ec08b920d6066d7071da7712" gracePeriod=30 Dec 10 09:53:02 crc kubenswrapper[4715]: I1210 09:53:02.275359 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c86fc9d-jwnlv" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerName="barbican-api" containerID="cri-o://44a195dfa862876a9a7e529edabaed0a7ac5415ff5d6db8d8ce8a825b5c92699" gracePeriod=30 Dec 10 09:53:02 crc kubenswrapper[4715]: I1210 09:53:02.606552 4715 generic.go:334] "Generic (PLEG): container finished" podID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerID="574f876f32cc798d7b9d401bb1023b5c1f7d1505ec08b920d6066d7071da7712" exitCode=143 Dec 10 09:53:02 crc kubenswrapper[4715]: I1210 09:53:02.606802 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86fc9d-jwnlv" event={"ID":"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2","Type":"ContainerDied","Data":"574f876f32cc798d7b9d401bb1023b5c1f7d1505ec08b920d6066d7071da7712"} Dec 10 09:53:03 crc kubenswrapper[4715]: I1210 09:53:03.967956 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 10 09:53:04 crc kubenswrapper[4715]: I1210 09:53:04.986390 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 10 09:53:05 crc kubenswrapper[4715]: I1210 09:53:05.373265 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-d9c5c96bf-n94nb" Dec 10 09:53:05 crc kubenswrapper[4715]: I1210 09:53:05.470160 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c86fc9d-jwnlv" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:33254->10.217.0.160:9311: read: connection reset by peer" Dec 10 09:53:05 crc kubenswrapper[4715]: I1210 09:53:05.470459 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c86fc9d-jwnlv" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:33270->10.217.0.160:9311: read: connection reset by peer" Dec 10 09:53:05 crc kubenswrapper[4715]: I1210 09:53:05.676843 4715 generic.go:334] "Generic (PLEG): container finished" podID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerID="44a195dfa862876a9a7e529edabaed0a7ac5415ff5d6db8d8ce8a825b5c92699" exitCode=0 Dec 10 09:53:05 crc kubenswrapper[4715]: I1210 09:53:05.676984 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86fc9d-jwnlv" event={"ID":"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2","Type":"ContainerDied","Data":"44a195dfa862876a9a7e529edabaed0a7ac5415ff5d6db8d8ce8a825b5c92699"} Dec 10 09:53:05 crc kubenswrapper[4715]: I1210 09:53:05.965669 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.005039 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-combined-ca-bundle\") pod \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.005078 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data-custom\") pod \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.005187 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkswn\" (UniqueName: \"kubernetes.io/projected/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-kube-api-access-jkswn\") pod \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.005213 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-logs\") pod \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.005231 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data\") pod \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\" (UID: \"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2\") " Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.012384 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" (UID: "7fbcf2f5-01d1-4607-8bd2-de5b5829afd2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.017350 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-logs" (OuterVolumeSpecName: "logs") pod "7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" (UID: "7fbcf2f5-01d1-4607-8bd2-de5b5829afd2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.024742 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-kube-api-access-jkswn" (OuterVolumeSpecName: "kube-api-access-jkswn") pod "7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" (UID: "7fbcf2f5-01d1-4607-8bd2-de5b5829afd2"). InnerVolumeSpecName "kube-api-access-jkswn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.049358 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" (UID: "7fbcf2f5-01d1-4607-8bd2-de5b5829afd2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.069601 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data" (OuterVolumeSpecName: "config-data") pod "7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" (UID: "7fbcf2f5-01d1-4607-8bd2-de5b5829afd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.106949 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.106983 4715 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.106994 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkswn\" (UniqueName: \"kubernetes.io/projected/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-kube-api-access-jkswn\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.107005 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.107014 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.691437 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86fc9d-jwnlv" event={"ID":"7fbcf2f5-01d1-4607-8bd2-de5b5829afd2","Type":"ContainerDied","Data":"306cfa120f83e0c8818bde62ea10d26d1cdf4e7b3c1ce0ea8fa82bdbcefc152d"} Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.691502 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c86fc9d-jwnlv" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.691772 4715 scope.go:117] "RemoveContainer" containerID="44a195dfa862876a9a7e529edabaed0a7ac5415ff5d6db8d8ce8a825b5c92699" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.723815 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c86fc9d-jwnlv"] Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.731241 4715 scope.go:117] "RemoveContainer" containerID="574f876f32cc798d7b9d401bb1023b5c1f7d1505ec08b920d6066d7071da7712" Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.733096 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6c86fc9d-jwnlv"] Dec 10 09:53:06 crc kubenswrapper[4715]: I1210 09:53:06.875475 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67f6cc7984-tshct" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.617573 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" path="/var/lib/kubelet/pods/7fbcf2f5-01d1-4607-8bd2-de5b5829afd2/volumes" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.900713 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6bcc74b4c5-ws76b"] Dec 10 09:53:07 crc kubenswrapper[4715]: E1210 09:53:07.901187 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerName="barbican-api-log" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.901205 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerName="barbican-api-log" Dec 10 09:53:07 crc kubenswrapper[4715]: E1210 09:53:07.901243 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerName="barbican-api" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.901249 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerName="barbican-api" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.901414 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerName="barbican-api" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.901434 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fbcf2f5-01d1-4607-8bd2-de5b5829afd2" containerName="barbican-api-log" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.902694 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.906377 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.906590 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.906764 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 10 09:53:07 crc kubenswrapper[4715]: I1210 09:53:07.913762 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6bcc74b4c5-ws76b"] Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.042557 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-public-tls-certs\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.042607 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-config-data\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.043460 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt269\" (UniqueName: \"kubernetes.io/projected/34a9b314-68fd-43d1-89c8-b55bf717f6df-kube-api-access-dt269\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.043518 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/34a9b314-68fd-43d1-89c8-b55bf717f6df-etc-swift\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.043537 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34a9b314-68fd-43d1-89c8-b55bf717f6df-run-httpd\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.043559 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-internal-tls-certs\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.043593 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-combined-ca-bundle\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.043622 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34a9b314-68fd-43d1-89c8-b55bf717f6df-log-httpd\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.145580 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/34a9b314-68fd-43d1-89c8-b55bf717f6df-etc-swift\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.145623 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34a9b314-68fd-43d1-89c8-b55bf717f6df-run-httpd\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.145647 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-internal-tls-certs\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.145680 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-combined-ca-bundle\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.145709 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34a9b314-68fd-43d1-89c8-b55bf717f6df-log-httpd\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.145749 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-public-tls-certs\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.145773 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-config-data\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.145847 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt269\" (UniqueName: \"kubernetes.io/projected/34a9b314-68fd-43d1-89c8-b55bf717f6df-kube-api-access-dt269\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.146207 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34a9b314-68fd-43d1-89c8-b55bf717f6df-run-httpd\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.147217 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34a9b314-68fd-43d1-89c8-b55bf717f6df-log-httpd\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.153018 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-internal-tls-certs\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.153048 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-combined-ca-bundle\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.153370 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-config-data\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.154157 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a9b314-68fd-43d1-89c8-b55bf717f6df-public-tls-certs\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.155348 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/34a9b314-68fd-43d1-89c8-b55bf717f6df-etc-swift\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.173942 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt269\" (UniqueName: \"kubernetes.io/projected/34a9b314-68fd-43d1-89c8-b55bf717f6df-kube-api-access-dt269\") pod \"swift-proxy-6bcc74b4c5-ws76b\" (UID: \"34a9b314-68fd-43d1-89c8-b55bf717f6df\") " pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.239559 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:08 crc kubenswrapper[4715]: I1210 09:53:08.977750 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6bcc74b4c5-ws76b"] Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.224480 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.552967 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.568105 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.570065 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.570428 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-npgsk" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.570729 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.574730 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.618662 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh6ll\" (UniqueName: \"kubernetes.io/projected/2843e1e0-1b99-45af-a422-e533ee493c78-kube-api-access-qh6ll\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.618758 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2843e1e0-1b99-45af-a422-e533ee493c78-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.618806 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2843e1e0-1b99-45af-a422-e533ee493c78-openstack-config\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.618854 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2843e1e0-1b99-45af-a422-e533ee493c78-openstack-config-secret\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.721012 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2843e1e0-1b99-45af-a422-e533ee493c78-openstack-config\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.721081 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2843e1e0-1b99-45af-a422-e533ee493c78-openstack-config-secret\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.721219 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh6ll\" (UniqueName: \"kubernetes.io/projected/2843e1e0-1b99-45af-a422-e533ee493c78-kube-api-access-qh6ll\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.721260 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2843e1e0-1b99-45af-a422-e533ee493c78-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.722130 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2843e1e0-1b99-45af-a422-e533ee493c78-openstack-config\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.725191 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2843e1e0-1b99-45af-a422-e533ee493c78-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.725551 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2843e1e0-1b99-45af-a422-e533ee493c78-openstack-config-secret\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.742292 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh6ll\" (UniqueName: \"kubernetes.io/projected/2843e1e0-1b99-45af-a422-e533ee493c78-kube-api-access-qh6ll\") pod \"openstackclient\" (UID: \"2843e1e0-1b99-45af-a422-e533ee493c78\") " pod="openstack/openstackclient" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.778184 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6bcc74b4c5-ws76b" event={"ID":"34a9b314-68fd-43d1-89c8-b55bf717f6df","Type":"ContainerStarted","Data":"b808a9dec354fdd6f4c38219231d0ca02adebe45b1b0767fc2e046eb7b6101a8"} Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.778234 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6bcc74b4c5-ws76b" event={"ID":"34a9b314-68fd-43d1-89c8-b55bf717f6df","Type":"ContainerStarted","Data":"9fe7d259223aa01594b8b9fb0f658b3222d11fa9b24f1ce0b35def39adb0583b"} Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.778251 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6bcc74b4c5-ws76b" event={"ID":"34a9b314-68fd-43d1-89c8-b55bf717f6df","Type":"ContainerStarted","Data":"ad8f9d55c676dc761825f549aad914d8a684f371e1f1cdad35f3cf4a745c8901"} Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.778367 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.800127 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6bcc74b4c5-ws76b" podStartSLOduration=2.800108733 podStartE2EDuration="2.800108733s" podCreationTimestamp="2025-12-10 09:53:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:53:09.79542649 +0000 UTC m=+1152.538972741" watchObservedRunningTime="2025-12-10 09:53:09.800108733 +0000 UTC m=+1152.543654984" Dec 10 09:53:09 crc kubenswrapper[4715]: I1210 09:53:09.898536 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 10 09:53:10 crc kubenswrapper[4715]: I1210 09:53:10.352354 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 10 09:53:10 crc kubenswrapper[4715]: W1210 09:53:10.361087 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2843e1e0_1b99_45af_a422_e533ee493c78.slice/crio-c899285ac2ad4b486c78d296dbf56b4421d36ef526728597ab26192c2526cfcd WatchSource:0}: Error finding container c899285ac2ad4b486c78d296dbf56b4421d36ef526728597ab26192c2526cfcd: Status 404 returned error can't find the container with id c899285ac2ad4b486c78d296dbf56b4421d36ef526728597ab26192c2526cfcd Dec 10 09:53:10 crc kubenswrapper[4715]: I1210 09:53:10.801175 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2843e1e0-1b99-45af-a422-e533ee493c78","Type":"ContainerStarted","Data":"c899285ac2ad4b486c78d296dbf56b4421d36ef526728597ab26192c2526cfcd"} Dec 10 09:53:10 crc kubenswrapper[4715]: I1210 09:53:10.801489 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:10 crc kubenswrapper[4715]: I1210 09:53:10.860048 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:10 crc kubenswrapper[4715]: I1210 09:53:10.860398 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="ceilometer-central-agent" containerID="cri-o://fdffd979d1ec3cb49ed1246f33350d6b6b6238dca028a1073765d92ce628d7cb" gracePeriod=30 Dec 10 09:53:10 crc kubenswrapper[4715]: I1210 09:53:10.860991 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="sg-core" containerID="cri-o://bb5112104ae2dcd87b19d76c289b8d38c5c89a6ab6e2f75f72729cf6a5a8062a" gracePeriod=30 Dec 10 09:53:10 crc kubenswrapper[4715]: I1210 09:53:10.861087 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="proxy-httpd" containerID="cri-o://bbdba2f1b665159bcef79064a0181735663c38eba05c61cea686e9ca34a9f6ac" gracePeriod=30 Dec 10 09:53:10 crc kubenswrapper[4715]: I1210 09:53:10.861192 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="ceilometer-notification-agent" containerID="cri-o://67ce57ae17678110c220b601a055667d6f51c6269cad420b147e48a7fbf99527" gracePeriod=30 Dec 10 09:53:10 crc kubenswrapper[4715]: I1210 09:53:10.965016 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.164:3000/\": read tcp 10.217.0.2:47168->10.217.0.164:3000: read: connection reset by peer" Dec 10 09:53:11 crc kubenswrapper[4715]: I1210 09:53:11.813069 4715 generic.go:334] "Generic (PLEG): container finished" podID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerID="bbdba2f1b665159bcef79064a0181735663c38eba05c61cea686e9ca34a9f6ac" exitCode=0 Dec 10 09:53:11 crc kubenswrapper[4715]: I1210 09:53:11.813448 4715 generic.go:334] "Generic (PLEG): container finished" podID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerID="bb5112104ae2dcd87b19d76c289b8d38c5c89a6ab6e2f75f72729cf6a5a8062a" exitCode=2 Dec 10 09:53:11 crc kubenswrapper[4715]: I1210 09:53:11.813462 4715 generic.go:334] "Generic (PLEG): container finished" podID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerID="fdffd979d1ec3cb49ed1246f33350d6b6b6238dca028a1073765d92ce628d7cb" exitCode=0 Dec 10 09:53:11 crc kubenswrapper[4715]: I1210 09:53:11.813161 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bacefd07-78b8-45cb-930b-5c4ec9fa9b95","Type":"ContainerDied","Data":"bbdba2f1b665159bcef79064a0181735663c38eba05c61cea686e9ca34a9f6ac"} Dec 10 09:53:11 crc kubenswrapper[4715]: I1210 09:53:11.813585 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bacefd07-78b8-45cb-930b-5c4ec9fa9b95","Type":"ContainerDied","Data":"bb5112104ae2dcd87b19d76c289b8d38c5c89a6ab6e2f75f72729cf6a5a8062a"} Dec 10 09:53:11 crc kubenswrapper[4715]: I1210 09:53:11.813607 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bacefd07-78b8-45cb-930b-5c4ec9fa9b95","Type":"ContainerDied","Data":"fdffd979d1ec3cb49ed1246f33350d6b6b6238dca028a1073765d92ce628d7cb"} Dec 10 09:53:15 crc kubenswrapper[4715]: I1210 09:53:15.848925 4715 generic.go:334] "Generic (PLEG): container finished" podID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerID="67ce57ae17678110c220b601a055667d6f51c6269cad420b147e48a7fbf99527" exitCode=0 Dec 10 09:53:15 crc kubenswrapper[4715]: I1210 09:53:15.849381 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bacefd07-78b8-45cb-930b-5c4ec9fa9b95","Type":"ContainerDied","Data":"67ce57ae17678110c220b601a055667d6f51c6269cad420b147e48a7fbf99527"} Dec 10 09:53:16 crc kubenswrapper[4715]: I1210 09:53:16.860534 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67f6cc7984-tshct" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 10 09:53:16 crc kubenswrapper[4715]: I1210 09:53:16.860631 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:53:17 crc kubenswrapper[4715]: I1210 09:53:17.713675 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:53:17 crc kubenswrapper[4715]: I1210 09:53:17.713744 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:53:18 crc kubenswrapper[4715]: I1210 09:53:18.262678 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:18 crc kubenswrapper[4715]: I1210 09:53:18.266786 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6bcc74b4c5-ws76b" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.164137 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.328810 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-log-httpd\") pod \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.328879 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-run-httpd\") pod \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.329035 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-scripts\") pod \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.329086 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-sg-core-conf-yaml\") pod \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.329133 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2ggl\" (UniqueName: \"kubernetes.io/projected/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-kube-api-access-h2ggl\") pod \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.329198 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-config-data\") pod \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.329259 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-combined-ca-bundle\") pod \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\" (UID: \"bacefd07-78b8-45cb-930b-5c4ec9fa9b95\") " Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.329319 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bacefd07-78b8-45cb-930b-5c4ec9fa9b95" (UID: "bacefd07-78b8-45cb-930b-5c4ec9fa9b95"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.329420 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bacefd07-78b8-45cb-930b-5c4ec9fa9b95" (UID: "bacefd07-78b8-45cb-930b-5c4ec9fa9b95"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.330096 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.330111 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.335537 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-kube-api-access-h2ggl" (OuterVolumeSpecName: "kube-api-access-h2ggl") pod "bacefd07-78b8-45cb-930b-5c4ec9fa9b95" (UID: "bacefd07-78b8-45cb-930b-5c4ec9fa9b95"). InnerVolumeSpecName "kube-api-access-h2ggl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.335558 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-scripts" (OuterVolumeSpecName: "scripts") pod "bacefd07-78b8-45cb-930b-5c4ec9fa9b95" (UID: "bacefd07-78b8-45cb-930b-5c4ec9fa9b95"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.379963 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bacefd07-78b8-45cb-930b-5c4ec9fa9b95" (UID: "bacefd07-78b8-45cb-930b-5c4ec9fa9b95"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.430082 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bacefd07-78b8-45cb-930b-5c4ec9fa9b95" (UID: "bacefd07-78b8-45cb-930b-5c4ec9fa9b95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.431741 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.431797 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.431820 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2ggl\" (UniqueName: \"kubernetes.io/projected/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-kube-api-access-h2ggl\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.431842 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.464261 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-config-data" (OuterVolumeSpecName: "config-data") pod "bacefd07-78b8-45cb-930b-5c4ec9fa9b95" (UID: "bacefd07-78b8-45cb-930b-5c4ec9fa9b95"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.533769 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bacefd07-78b8-45cb-930b-5c4ec9fa9b95-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.888992 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bacefd07-78b8-45cb-930b-5c4ec9fa9b95","Type":"ContainerDied","Data":"01eada311fa18bba2b676fad1667ca814b8f7af89bf8f8762f840e14863420ca"} Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.889067 4715 scope.go:117] "RemoveContainer" containerID="bbdba2f1b665159bcef79064a0181735663c38eba05c61cea686e9ca34a9f6ac" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.889324 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.894183 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2843e1e0-1b99-45af-a422-e533ee493c78","Type":"ContainerStarted","Data":"74476b13c3b7e46e284b3d69b9963db117b4790ca0105a8f3331e02d10308177"} Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.923125 4715 scope.go:117] "RemoveContainer" containerID="bb5112104ae2dcd87b19d76c289b8d38c5c89a6ab6e2f75f72729cf6a5a8062a" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.924593 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.402542043 podStartE2EDuration="10.924577037s" podCreationTimestamp="2025-12-10 09:53:09 +0000 UTC" firstStartedPulling="2025-12-10 09:53:10.363627804 +0000 UTC m=+1153.107174055" lastFinishedPulling="2025-12-10 09:53:18.885662798 +0000 UTC m=+1161.629209049" observedRunningTime="2025-12-10 09:53:19.91692678 +0000 UTC m=+1162.660473031" watchObservedRunningTime="2025-12-10 09:53:19.924577037 +0000 UTC m=+1162.668123288" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.942260 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.960136 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.964427 4715 scope.go:117] "RemoveContainer" containerID="67ce57ae17678110c220b601a055667d6f51c6269cad420b147e48a7fbf99527" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.975008 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:19 crc kubenswrapper[4715]: E1210 09:53:19.975483 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="ceilometer-central-agent" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.975508 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="ceilometer-central-agent" Dec 10 09:53:19 crc kubenswrapper[4715]: E1210 09:53:19.975524 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="proxy-httpd" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.975533 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="proxy-httpd" Dec 10 09:53:19 crc kubenswrapper[4715]: E1210 09:53:19.975578 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="sg-core" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.975586 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="sg-core" Dec 10 09:53:19 crc kubenswrapper[4715]: E1210 09:53:19.975600 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="ceilometer-notification-agent" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.975610 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="ceilometer-notification-agent" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.975811 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="sg-core" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.975835 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="proxy-httpd" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.975856 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="ceilometer-notification-agent" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.975870 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" containerName="ceilometer-central-agent" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.977938 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.980371 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.984069 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 09:53:19 crc kubenswrapper[4715]: I1210 09:53:19.998445 4715 scope.go:117] "RemoveContainer" containerID="fdffd979d1ec3cb49ed1246f33350d6b6b6238dca028a1073765d92ce628d7cb" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.001801 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.149019 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-scripts\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.149139 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-config-data\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.149193 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-run-httpd\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.149256 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wh8v\" (UniqueName: \"kubernetes.io/projected/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-kube-api-access-6wh8v\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.149286 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-log-httpd\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.149310 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.149368 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.251350 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.251420 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-scripts\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.251504 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-config-data\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.251545 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-run-httpd\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.251615 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wh8v\" (UniqueName: \"kubernetes.io/projected/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-kube-api-access-6wh8v\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.251655 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-log-httpd\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.251682 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.253197 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-run-httpd\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.255569 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-log-httpd\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.264111 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-scripts\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.264537 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.285163 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-config-data\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.285190 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.298340 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wh8v\" (UniqueName: \"kubernetes.io/projected/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-kube-api-access-6wh8v\") pod \"ceilometer-0\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.305759 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:20 crc kubenswrapper[4715]: W1210 09:53:20.819814 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20acc89a_f0b2_40d9_a615_11c15f5f3e7f.slice/crio-edc5db340778b156c65ce4e18edbc868f7949f4db06cf64f35624a890c0d6412 WatchSource:0}: Error finding container edc5db340778b156c65ce4e18edbc868f7949f4db06cf64f35624a890c0d6412: Status 404 returned error can't find the container with id edc5db340778b156c65ce4e18edbc868f7949f4db06cf64f35624a890c0d6412 Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.820592 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:20 crc kubenswrapper[4715]: I1210 09:53:20.908388 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20acc89a-f0b2-40d9-a615-11c15f5f3e7f","Type":"ContainerStarted","Data":"edc5db340778b156c65ce4e18edbc868f7949f4db06cf64f35624a890c0d6412"} Dec 10 09:53:21 crc kubenswrapper[4715]: I1210 09:53:21.616614 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bacefd07-78b8-45cb-930b-5c4ec9fa9b95" path="/var/lib/kubelet/pods/bacefd07-78b8-45cb-930b-5c4ec9fa9b95/volumes" Dec 10 09:53:21 crc kubenswrapper[4715]: I1210 09:53:21.938141 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20acc89a-f0b2-40d9-a615-11c15f5f3e7f","Type":"ContainerStarted","Data":"90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894"} Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.175650 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.246197 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.246732 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerName="glance-log" containerID="cri-o://8ff58dd4c00af545e2be2e5ebc3e656b8e492796b8d42a11c5a8b4dbbb62373b" gracePeriod=30 Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.247643 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerName="glance-httpd" containerID="cri-o://51e47c5a22fe27694a01c0e2b0c68edc82cd7c5c052747e4d8a03a8164dbf545" gracePeriod=30 Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.793411 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.947832 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-config-data\") pod \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.947878 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-scripts\") pod \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.947935 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-combined-ca-bundle\") pod \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.948010 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-logs\") pod \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.948052 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-secret-key\") pod \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.948096 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-tls-certs\") pod \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.948139 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7czs\" (UniqueName: \"kubernetes.io/projected/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-kube-api-access-n7czs\") pod \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\" (UID: \"5a1c5abd-cfc5-427f-a378-cef6d31deb0e\") " Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.948452 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-logs" (OuterVolumeSpecName: "logs") pod "5a1c5abd-cfc5-427f-a378-cef6d31deb0e" (UID: "5a1c5abd-cfc5-427f-a378-cef6d31deb0e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.948979 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.955044 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5a1c5abd-cfc5-427f-a378-cef6d31deb0e" (UID: "5a1c5abd-cfc5-427f-a378-cef6d31deb0e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.955140 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-kube-api-access-n7czs" (OuterVolumeSpecName: "kube-api-access-n7czs") pod "5a1c5abd-cfc5-427f-a378-cef6d31deb0e" (UID: "5a1c5abd-cfc5-427f-a378-cef6d31deb0e"). InnerVolumeSpecName "kube-api-access-n7czs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.975515 4715 generic.go:334] "Generic (PLEG): container finished" podID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerID="c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171" exitCode=137 Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.975641 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f6cc7984-tshct" event={"ID":"5a1c5abd-cfc5-427f-a378-cef6d31deb0e","Type":"ContainerDied","Data":"c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171"} Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.975675 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f6cc7984-tshct" event={"ID":"5a1c5abd-cfc5-427f-a378-cef6d31deb0e","Type":"ContainerDied","Data":"1ece62e33ef77421cffb2e5b7d6c6ffa3f2dbacaca2f6365b9fbc5b4bb8fe542"} Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.975697 4715 scope.go:117] "RemoveContainer" containerID="e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d" Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.975870 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67f6cc7984-tshct" Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.989097 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a1c5abd-cfc5-427f-a378-cef6d31deb0e" (UID: "5a1c5abd-cfc5-427f-a378-cef6d31deb0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.989321 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20acc89a-f0b2-40d9-a615-11c15f5f3e7f","Type":"ContainerStarted","Data":"9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c"} Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.992088 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-config-data" (OuterVolumeSpecName: "config-data") pod "5a1c5abd-cfc5-427f-a378-cef6d31deb0e" (UID: "5a1c5abd-cfc5-427f-a378-cef6d31deb0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.997478 4715 generic.go:334] "Generic (PLEG): container finished" podID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerID="8ff58dd4c00af545e2be2e5ebc3e656b8e492796b8d42a11c5a8b4dbbb62373b" exitCode=143 Dec 10 09:53:22 crc kubenswrapper[4715]: I1210 09:53:22.997539 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"28610e7f-ab63-4b49-b9d5-73b17937f1b6","Type":"ContainerDied","Data":"8ff58dd4c00af545e2be2e5ebc3e656b8e492796b8d42a11c5a8b4dbbb62373b"} Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.078500 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-scripts" (OuterVolumeSpecName: "scripts") pod "5a1c5abd-cfc5-427f-a378-cef6d31deb0e" (UID: "5a1c5abd-cfc5-427f-a378-cef6d31deb0e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.086088 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.086141 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7czs\" (UniqueName: \"kubernetes.io/projected/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-kube-api-access-n7czs\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.086158 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.086169 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.086179 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.098007 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "5a1c5abd-cfc5-427f-a378-cef6d31deb0e" (UID: "5a1c5abd-cfc5-427f-a378-cef6d31deb0e"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.188958 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c5abd-cfc5-427f-a378-cef6d31deb0e-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.267489 4715 scope.go:117] "RemoveContainer" containerID="c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.299065 4715 scope.go:117] "RemoveContainer" containerID="e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d" Dec 10 09:53:23 crc kubenswrapper[4715]: E1210 09:53:23.301588 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d\": container with ID starting with e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d not found: ID does not exist" containerID="e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.301644 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d"} err="failed to get container status \"e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d\": rpc error: code = NotFound desc = could not find container \"e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d\": container with ID starting with e9604e25e74f0132134e469b6c444db80c29dad8b23d9253d187c88426b92c3d not found: ID does not exist" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.301667 4715 scope.go:117] "RemoveContainer" containerID="c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171" Dec 10 09:53:23 crc kubenswrapper[4715]: E1210 09:53:23.301956 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171\": container with ID starting with c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171 not found: ID does not exist" containerID="c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.301999 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171"} err="failed to get container status \"c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171\": rpc error: code = NotFound desc = could not find container \"c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171\": container with ID starting with c6c9a76435f314ed019ca745ad740f72d10c009b9ea8bfbc50359880bb325171 not found: ID does not exist" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.335801 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67f6cc7984-tshct"] Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.346339 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-67f6cc7984-tshct"] Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.618760 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" path="/var/lib/kubelet/pods/5a1c5abd-cfc5-427f-a378-cef6d31deb0e/volumes" Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.630783 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.631096 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="207b4574-9a00-43c1-9d73-24e23343f35c" containerName="glance-log" containerID="cri-o://d3d0255b6696448d82f647922157d65706d9eeb197b2dab337d7f96103ad86b1" gracePeriod=30 Dec 10 09:53:23 crc kubenswrapper[4715]: I1210 09:53:23.631202 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="207b4574-9a00-43c1-9d73-24e23343f35c" containerName="glance-httpd" containerID="cri-o://c37868688530d331997fbb4762beee580a5d5cba9d10c530844a74cea4b527c7" gracePeriod=30 Dec 10 09:53:24 crc kubenswrapper[4715]: I1210 09:53:24.069122 4715 generic.go:334] "Generic (PLEG): container finished" podID="207b4574-9a00-43c1-9d73-24e23343f35c" containerID="d3d0255b6696448d82f647922157d65706d9eeb197b2dab337d7f96103ad86b1" exitCode=143 Dec 10 09:53:24 crc kubenswrapper[4715]: I1210 09:53:24.069345 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"207b4574-9a00-43c1-9d73-24e23343f35c","Type":"ContainerDied","Data":"d3d0255b6696448d82f647922157d65706d9eeb197b2dab337d7f96103ad86b1"} Dec 10 09:53:24 crc kubenswrapper[4715]: I1210 09:53:24.088476 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20acc89a-f0b2-40d9-a615-11c15f5f3e7f","Type":"ContainerStarted","Data":"ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec"} Dec 10 09:53:25 crc kubenswrapper[4715]: I1210 09:53:25.534642 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:49874->10.217.0.154:9292: read: connection reset by peer" Dec 10 09:53:25 crc kubenswrapper[4715]: I1210 09:53:25.534659 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:49862->10.217.0.154:9292: read: connection reset by peer" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.110796 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20acc89a-f0b2-40d9-a615-11c15f5f3e7f","Type":"ContainerStarted","Data":"e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c"} Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.111446 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="ceilometer-central-agent" containerID="cri-o://90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894" gracePeriod=30 Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.111526 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.111951 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="proxy-httpd" containerID="cri-o://e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c" gracePeriod=30 Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.112000 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="ceilometer-notification-agent" containerID="cri-o://9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c" gracePeriod=30 Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.112036 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="sg-core" containerID="cri-o://ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec" gracePeriod=30 Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.124044 4715 generic.go:334] "Generic (PLEG): container finished" podID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerID="51e47c5a22fe27694a01c0e2b0c68edc82cd7c5c052747e4d8a03a8164dbf545" exitCode=0 Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.124093 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"28610e7f-ab63-4b49-b9d5-73b17937f1b6","Type":"ContainerDied","Data":"51e47c5a22fe27694a01c0e2b0c68edc82cd7c5c052747e4d8a03a8164dbf545"} Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.124122 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"28610e7f-ab63-4b49-b9d5-73b17937f1b6","Type":"ContainerDied","Data":"c42ac83d795f3d8d04ef9effa327a8d30a5066d09fa57a026b9bbb64223a3f89"} Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.124135 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c42ac83d795f3d8d04ef9effa327a8d30a5066d09fa57a026b9bbb64223a3f89" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.167616 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.897913718 podStartE2EDuration="7.167591733s" podCreationTimestamp="2025-12-10 09:53:19 +0000 UTC" firstStartedPulling="2025-12-10 09:53:20.822331459 +0000 UTC m=+1163.565877710" lastFinishedPulling="2025-12-10 09:53:25.092009474 +0000 UTC m=+1167.835555725" observedRunningTime="2025-12-10 09:53:26.157102534 +0000 UTC m=+1168.900648785" watchObservedRunningTime="2025-12-10 09:53:26.167591733 +0000 UTC m=+1168.911137984" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.295082 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.343088 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-httpd-run\") pod \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.343246 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-scripts\") pod \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.343305 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2z8sm\" (UniqueName: \"kubernetes.io/projected/28610e7f-ab63-4b49-b9d5-73b17937f1b6-kube-api-access-2z8sm\") pod \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.343478 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-public-tls-certs\") pod \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.343680 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-logs\") pod \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.343713 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-config-data\") pod \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.343759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.343798 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-combined-ca-bundle\") pod \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\" (UID: \"28610e7f-ab63-4b49-b9d5-73b17937f1b6\") " Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.343873 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "28610e7f-ab63-4b49-b9d5-73b17937f1b6" (UID: "28610e7f-ab63-4b49-b9d5-73b17937f1b6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.344158 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-logs" (OuterVolumeSpecName: "logs") pod "28610e7f-ab63-4b49-b9d5-73b17937f1b6" (UID: "28610e7f-ab63-4b49-b9d5-73b17937f1b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.344856 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.344888 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28610e7f-ab63-4b49-b9d5-73b17937f1b6-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.364156 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28610e7f-ab63-4b49-b9d5-73b17937f1b6-kube-api-access-2z8sm" (OuterVolumeSpecName: "kube-api-access-2z8sm") pod "28610e7f-ab63-4b49-b9d5-73b17937f1b6" (UID: "28610e7f-ab63-4b49-b9d5-73b17937f1b6"). InnerVolumeSpecName "kube-api-access-2z8sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.384159 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-scripts" (OuterVolumeSpecName: "scripts") pod "28610e7f-ab63-4b49-b9d5-73b17937f1b6" (UID: "28610e7f-ab63-4b49-b9d5-73b17937f1b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.384283 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "28610e7f-ab63-4b49-b9d5-73b17937f1b6" (UID: "28610e7f-ab63-4b49-b9d5-73b17937f1b6"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.450869 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.450939 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.450956 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2z8sm\" (UniqueName: \"kubernetes.io/projected/28610e7f-ab63-4b49-b9d5-73b17937f1b6-kube-api-access-2z8sm\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.496116 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28610e7f-ab63-4b49-b9d5-73b17937f1b6" (UID: "28610e7f-ab63-4b49-b9d5-73b17937f1b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.501207 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-config-data" (OuterVolumeSpecName: "config-data") pod "28610e7f-ab63-4b49-b9d5-73b17937f1b6" (UID: "28610e7f-ab63-4b49-b9d5-73b17937f1b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.531572 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "28610e7f-ab63-4b49-b9d5-73b17937f1b6" (UID: "28610e7f-ab63-4b49-b9d5-73b17937f1b6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.553398 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.554566 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.554660 4715 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28610e7f-ab63-4b49-b9d5-73b17937f1b6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.560303 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 10 09:53:26 crc kubenswrapper[4715]: I1210 09:53:26.657089 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.137712 4715 generic.go:334] "Generic (PLEG): container finished" podID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerID="e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c" exitCode=0 Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.138115 4715 generic.go:334] "Generic (PLEG): container finished" podID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerID="ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec" exitCode=2 Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.138127 4715 generic.go:334] "Generic (PLEG): container finished" podID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerID="9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c" exitCode=0 Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.137775 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20acc89a-f0b2-40d9-a615-11c15f5f3e7f","Type":"ContainerDied","Data":"e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c"} Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.138264 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20acc89a-f0b2-40d9-a615-11c15f5f3e7f","Type":"ContainerDied","Data":"ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec"} Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.138294 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20acc89a-f0b2-40d9-a615-11c15f5f3e7f","Type":"ContainerDied","Data":"9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c"} Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.140479 4715 generic.go:334] "Generic (PLEG): container finished" podID="207b4574-9a00-43c1-9d73-24e23343f35c" containerID="c37868688530d331997fbb4762beee580a5d5cba9d10c530844a74cea4b527c7" exitCode=0 Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.140548 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"207b4574-9a00-43c1-9d73-24e23343f35c","Type":"ContainerDied","Data":"c37868688530d331997fbb4762beee580a5d5cba9d10c530844a74cea4b527c7"} Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.140590 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.367374 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.402237 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.455901 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:53:27 crc kubenswrapper[4715]: E1210 09:53:27.456565 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon-log" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.456589 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon-log" Dec 10 09:53:27 crc kubenswrapper[4715]: E1210 09:53:27.456605 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerName="glance-httpd" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.456615 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerName="glance-httpd" Dec 10 09:53:27 crc kubenswrapper[4715]: E1210 09:53:27.456685 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerName="glance-log" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.456697 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerName="glance-log" Dec 10 09:53:27 crc kubenswrapper[4715]: E1210 09:53:27.456713 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.456724 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.457111 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerName="glance-httpd" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.457146 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.457191 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1c5abd-cfc5-427f-a378-cef6d31deb0e" containerName="horizon-log" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.457223 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" containerName="glance-log" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.458856 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.462964 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.463161 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.466339 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.815639 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.815716 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-scripts\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.815743 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.815798 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.815821 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-logs\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.815859 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-config-data\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.815884 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.815906 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87ncb\" (UniqueName: \"kubernetes.io/projected/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-kube-api-access-87ncb\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.855624 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.856499 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28610e7f-ab63-4b49-b9d5-73b17937f1b6" path="/var/lib/kubelet/pods/28610e7f-ab63-4b49-b9d5-73b17937f1b6/volumes" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.917365 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.917703 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-logs\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.917744 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-config-data\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.917778 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.917815 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87ncb\" (UniqueName: \"kubernetes.io/projected/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-kube-api-access-87ncb\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.917999 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.918063 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-scripts\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.918099 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.918344 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-logs\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.918737 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.918748 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.927183 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.932401 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-config-data\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.935837 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-scripts\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.936154 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.940167 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87ncb\" (UniqueName: \"kubernetes.io/projected/92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac-kube-api-access-87ncb\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:27 crc kubenswrapper[4715]: I1210 09:53:27.967545 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac\") " pod="openstack/glance-default-external-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.020201 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-httpd-run\") pod \"207b4574-9a00-43c1-9d73-24e23343f35c\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.020377 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzbp8\" (UniqueName: \"kubernetes.io/projected/207b4574-9a00-43c1-9d73-24e23343f35c-kube-api-access-vzbp8\") pod \"207b4574-9a00-43c1-9d73-24e23343f35c\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.020410 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"207b4574-9a00-43c1-9d73-24e23343f35c\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.020430 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-logs\") pod \"207b4574-9a00-43c1-9d73-24e23343f35c\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.020660 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-scripts\") pod \"207b4574-9a00-43c1-9d73-24e23343f35c\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.020721 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-combined-ca-bundle\") pod \"207b4574-9a00-43c1-9d73-24e23343f35c\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.020758 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-config-data\") pod \"207b4574-9a00-43c1-9d73-24e23343f35c\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.020831 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-internal-tls-certs\") pod \"207b4574-9a00-43c1-9d73-24e23343f35c\" (UID: \"207b4574-9a00-43c1-9d73-24e23343f35c\") " Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.021087 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "207b4574-9a00-43c1-9d73-24e23343f35c" (UID: "207b4574-9a00-43c1-9d73-24e23343f35c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.021160 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-logs" (OuterVolumeSpecName: "logs") pod "207b4574-9a00-43c1-9d73-24e23343f35c" (UID: "207b4574-9a00-43c1-9d73-24e23343f35c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.021512 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.021545 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/207b4574-9a00-43c1-9d73-24e23343f35c-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.027029 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-scripts" (OuterVolumeSpecName: "scripts") pod "207b4574-9a00-43c1-9d73-24e23343f35c" (UID: "207b4574-9a00-43c1-9d73-24e23343f35c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.030136 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "207b4574-9a00-43c1-9d73-24e23343f35c" (UID: "207b4574-9a00-43c1-9d73-24e23343f35c"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.039819 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/207b4574-9a00-43c1-9d73-24e23343f35c-kube-api-access-vzbp8" (OuterVolumeSpecName: "kube-api-access-vzbp8") pod "207b4574-9a00-43c1-9d73-24e23343f35c" (UID: "207b4574-9a00-43c1-9d73-24e23343f35c"). InnerVolumeSpecName "kube-api-access-vzbp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.077360 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "207b4574-9a00-43c1-9d73-24e23343f35c" (UID: "207b4574-9a00-43c1-9d73-24e23343f35c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.088577 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.094170 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-config-data" (OuterVolumeSpecName: "config-data") pod "207b4574-9a00-43c1-9d73-24e23343f35c" (UID: "207b4574-9a00-43c1-9d73-24e23343f35c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.120684 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "207b4574-9a00-43c1-9d73-24e23343f35c" (UID: "207b4574-9a00-43c1-9d73-24e23343f35c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.123145 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzbp8\" (UniqueName: \"kubernetes.io/projected/207b4574-9a00-43c1-9d73-24e23343f35c-kube-api-access-vzbp8\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.123193 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.123209 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.123222 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.123233 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.123240 4715 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/207b4574-9a00-43c1-9d73-24e23343f35c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.182053 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"207b4574-9a00-43c1-9d73-24e23343f35c","Type":"ContainerDied","Data":"e3e5c7590b66d8da8450aeb403df0e3793c4bce42848ff430c0ae650fe038bf5"} Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.182140 4715 scope.go:117] "RemoveContainer" containerID="c37868688530d331997fbb4762beee580a5d5cba9d10c530844a74cea4b527c7" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.182705 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.195001 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.217269 4715 scope.go:117] "RemoveContainer" containerID="d3d0255b6696448d82f647922157d65706d9eeb197b2dab337d7f96103ad86b1" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.230515 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.255728 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.280979 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.295178 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:53:28 crc kubenswrapper[4715]: E1210 09:53:28.295550 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207b4574-9a00-43c1-9d73-24e23343f35c" containerName="glance-httpd" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.295562 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="207b4574-9a00-43c1-9d73-24e23343f35c" containerName="glance-httpd" Dec 10 09:53:28 crc kubenswrapper[4715]: E1210 09:53:28.295603 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207b4574-9a00-43c1-9d73-24e23343f35c" containerName="glance-log" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.295609 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="207b4574-9a00-43c1-9d73-24e23343f35c" containerName="glance-log" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.295770 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="207b4574-9a00-43c1-9d73-24e23343f35c" containerName="glance-httpd" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.295783 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="207b4574-9a00-43c1-9d73-24e23343f35c" containerName="glance-log" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.300026 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.303455 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.303630 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.304777 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.615158 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ba040a24-bae6-4627-a934-39d616b57c76-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.615425 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.615478 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.615664 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.615685 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5pzc\" (UniqueName: \"kubernetes.io/projected/ba040a24-bae6-4627-a934-39d616b57c76-kube-api-access-g5pzc\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.615709 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.615749 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba040a24-bae6-4627-a934-39d616b57c76-logs\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.615775 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.717950 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5pzc\" (UniqueName: \"kubernetes.io/projected/ba040a24-bae6-4627-a934-39d616b57c76-kube-api-access-g5pzc\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.718008 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.718036 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.718098 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba040a24-bae6-4627-a934-39d616b57c76-logs\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.718136 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.718185 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ba040a24-bae6-4627-a934-39d616b57c76-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.718236 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.718258 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.720137 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba040a24-bae6-4627-a934-39d616b57c76-logs\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.720189 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ba040a24-bae6-4627-a934-39d616b57c76-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.720333 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.729012 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.734395 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.742725 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5pzc\" (UniqueName: \"kubernetes.io/projected/ba040a24-bae6-4627-a934-39d616b57c76-kube-api-access-g5pzc\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.745522 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.758449 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba040a24-bae6-4627-a934-39d616b57c76-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.771198 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ba040a24-bae6-4627-a934-39d616b57c76\") " pod="openstack/glance-default-internal-api-0" Dec 10 09:53:28 crc kubenswrapper[4715]: I1210 09:53:28.903167 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:29 crc kubenswrapper[4715]: I1210 09:53:29.146630 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 09:53:29 crc kubenswrapper[4715]: I1210 09:53:29.284540 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac","Type":"ContainerStarted","Data":"71e8ef44e81ddf44d94eba6a6d916f87a241c957512a4b265269ce79a1b746c7"} Dec 10 09:53:29 crc kubenswrapper[4715]: I1210 09:53:29.618613 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="207b4574-9a00-43c1-9d73-24e23343f35c" path="/var/lib/kubelet/pods/207b4574-9a00-43c1-9d73-24e23343f35c/volumes" Dec 10 09:53:29 crc kubenswrapper[4715]: I1210 09:53:29.830799 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 09:53:30 crc kubenswrapper[4715]: I1210 09:53:30.322588 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac","Type":"ContainerStarted","Data":"f927687c9048a9e19a375d8e5f1103a8761f8ead6d4cbe1f96ea916aa6543a82"} Dec 10 09:53:30 crc kubenswrapper[4715]: I1210 09:53:30.323843 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ba040a24-bae6-4627-a934-39d616b57c76","Type":"ContainerStarted","Data":"b8a5b3577741e78e13efb051c4add54b026cb23f1bf43a7516d7e30ecc791df5"} Dec 10 09:53:31 crc kubenswrapper[4715]: I1210 09:53:31.343379 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac","Type":"ContainerStarted","Data":"d2107d794663112a8ded432e82ce3ad47a2a939a4f14e85cbe486abfbc779997"} Dec 10 09:53:31 crc kubenswrapper[4715]: I1210 09:53:31.347246 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ba040a24-bae6-4627-a934-39d616b57c76","Type":"ContainerStarted","Data":"13d092852b8ddbfa9f5db7262027e431d2854a36036d8c38eb983349d3ca8c2d"} Dec 10 09:53:32 crc kubenswrapper[4715]: I1210 09:53:32.358956 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ba040a24-bae6-4627-a934-39d616b57c76","Type":"ContainerStarted","Data":"e3b0d480b070a7322f0b27b1caf814a44ce365ffc4e08fdb3824d18a76655015"} Dec 10 09:53:32 crc kubenswrapper[4715]: I1210 09:53:32.381116 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.381095396 podStartE2EDuration="5.381095396s" podCreationTimestamp="2025-12-10 09:53:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:53:31.366107685 +0000 UTC m=+1174.109653936" watchObservedRunningTime="2025-12-10 09:53:32.381095396 +0000 UTC m=+1175.124641657" Dec 10 09:53:32 crc kubenswrapper[4715]: I1210 09:53:32.386158 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.386136359 podStartE2EDuration="4.386136359s" podCreationTimestamp="2025-12-10 09:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:53:32.378572634 +0000 UTC m=+1175.122118905" watchObservedRunningTime="2025-12-10 09:53:32.386136359 +0000 UTC m=+1175.129682610" Dec 10 09:53:32 crc kubenswrapper[4715]: I1210 09:53:32.821959 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.012237 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-config-data\") pod \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.012334 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wh8v\" (UniqueName: \"kubernetes.io/projected/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-kube-api-access-6wh8v\") pod \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.012381 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-combined-ca-bundle\") pod \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.012478 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-sg-core-conf-yaml\") pod \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.012530 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-run-httpd\") pod \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.012609 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-log-httpd\") pod \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.012651 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-scripts\") pod \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\" (UID: \"20acc89a-f0b2-40d9-a615-11c15f5f3e7f\") " Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.016377 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "20acc89a-f0b2-40d9-a615-11c15f5f3e7f" (UID: "20acc89a-f0b2-40d9-a615-11c15f5f3e7f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.016712 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "20acc89a-f0b2-40d9-a615-11c15f5f3e7f" (UID: "20acc89a-f0b2-40d9-a615-11c15f5f3e7f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.019687 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-scripts" (OuterVolumeSpecName: "scripts") pod "20acc89a-f0b2-40d9-a615-11c15f5f3e7f" (UID: "20acc89a-f0b2-40d9-a615-11c15f5f3e7f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.048184 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "20acc89a-f0b2-40d9-a615-11c15f5f3e7f" (UID: "20acc89a-f0b2-40d9-a615-11c15f5f3e7f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.057146 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-kube-api-access-6wh8v" (OuterVolumeSpecName: "kube-api-access-6wh8v") pod "20acc89a-f0b2-40d9-a615-11c15f5f3e7f" (UID: "20acc89a-f0b2-40d9-a615-11c15f5f3e7f"). InnerVolumeSpecName "kube-api-access-6wh8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.114665 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wh8v\" (UniqueName: \"kubernetes.io/projected/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-kube-api-access-6wh8v\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.114700 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.114712 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.114724 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.114736 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.128388 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20acc89a-f0b2-40d9-a615-11c15f5f3e7f" (UID: "20acc89a-f0b2-40d9-a615-11c15f5f3e7f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.148816 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-config-data" (OuterVolumeSpecName: "config-data") pod "20acc89a-f0b2-40d9-a615-11c15f5f3e7f" (UID: "20acc89a-f0b2-40d9-a615-11c15f5f3e7f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.217333 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.217405 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20acc89a-f0b2-40d9-a615-11c15f5f3e7f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.371561 4715 generic.go:334] "Generic (PLEG): container finished" podID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerID="90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894" exitCode=0 Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.371609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20acc89a-f0b2-40d9-a615-11c15f5f3e7f","Type":"ContainerDied","Data":"90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894"} Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.371665 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20acc89a-f0b2-40d9-a615-11c15f5f3e7f","Type":"ContainerDied","Data":"edc5db340778b156c65ce4e18edbc868f7949f4db06cf64f35624a890c0d6412"} Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.371710 4715 scope.go:117] "RemoveContainer" containerID="e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.371728 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.397069 4715 scope.go:117] "RemoveContainer" containerID="ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.418117 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.427674 4715 scope.go:117] "RemoveContainer" containerID="9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.436085 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.446892 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:33 crc kubenswrapper[4715]: E1210 09:53:33.447368 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="sg-core" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.447390 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="sg-core" Dec 10 09:53:33 crc kubenswrapper[4715]: E1210 09:53:33.447416 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="ceilometer-notification-agent" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.447425 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="ceilometer-notification-agent" Dec 10 09:53:33 crc kubenswrapper[4715]: E1210 09:53:33.447448 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="ceilometer-central-agent" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.447457 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="ceilometer-central-agent" Dec 10 09:53:33 crc kubenswrapper[4715]: E1210 09:53:33.447495 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="proxy-httpd" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.447504 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="proxy-httpd" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.447713 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="ceilometer-notification-agent" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.447735 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="ceilometer-central-agent" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.447760 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="sg-core" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.447779 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" containerName="proxy-httpd" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.449812 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.453251 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.455369 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.467698 4715 scope.go:117] "RemoveContainer" containerID="90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.479589 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.544846 4715 scope.go:117] "RemoveContainer" containerID="e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c" Dec 10 09:53:33 crc kubenswrapper[4715]: E1210 09:53:33.545338 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c\": container with ID starting with e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c not found: ID does not exist" containerID="e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.545379 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c"} err="failed to get container status \"e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c\": rpc error: code = NotFound desc = could not find container \"e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c\": container with ID starting with e80ff930a465540c072b6e35276ebb365eb65180746cdfb62b633d43e2d3802c not found: ID does not exist" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.545406 4715 scope.go:117] "RemoveContainer" containerID="ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec" Dec 10 09:53:33 crc kubenswrapper[4715]: E1210 09:53:33.545802 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec\": container with ID starting with ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec not found: ID does not exist" containerID="ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.545864 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec"} err="failed to get container status \"ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec\": rpc error: code = NotFound desc = could not find container \"ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec\": container with ID starting with ca4cc1fd12b37c96bd40badaa7f5c8220aa5178a74d3f8df7e9f6d966032d3ec not found: ID does not exist" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.545892 4715 scope.go:117] "RemoveContainer" containerID="9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c" Dec 10 09:53:33 crc kubenswrapper[4715]: E1210 09:53:33.546238 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c\": container with ID starting with 9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c not found: ID does not exist" containerID="9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.546265 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c"} err="failed to get container status \"9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c\": rpc error: code = NotFound desc = could not find container \"9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c\": container with ID starting with 9d95bdf136bb7e35bcb16dbb2f56c9490a425eb6fe8451d253c6f590c21c2e1c not found: ID does not exist" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.546283 4715 scope.go:117] "RemoveContainer" containerID="90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894" Dec 10 09:53:33 crc kubenswrapper[4715]: E1210 09:53:33.547067 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894\": container with ID starting with 90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894 not found: ID does not exist" containerID="90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.547094 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894"} err="failed to get container status \"90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894\": rpc error: code = NotFound desc = could not find container \"90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894\": container with ID starting with 90e8dd3a7ab4c254f0225bca8fb23b448288839f99bd5480b07a2ded81427894 not found: ID does not exist" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.617108 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20acc89a-f0b2-40d9-a615-11c15f5f3e7f" path="/var/lib/kubelet/pods/20acc89a-f0b2-40d9-a615-11c15f5f3e7f/volumes" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.624870 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.624954 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.625009 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-config-data\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.625032 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-run-httpd\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.625210 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-log-httpd\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.625419 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfrsf\" (UniqueName: \"kubernetes.io/projected/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-kube-api-access-pfrsf\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.625701 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-scripts\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.727775 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.727844 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.727905 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-config-data\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.727974 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-run-httpd\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.728525 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-run-httpd\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.728675 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-log-httpd\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.728767 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfrsf\" (UniqueName: \"kubernetes.io/projected/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-kube-api-access-pfrsf\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.728837 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-scripts\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.729081 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-log-httpd\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.732805 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.733261 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.733736 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-scripts\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.733768 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-config-data\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.747774 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfrsf\" (UniqueName: \"kubernetes.io/projected/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-kube-api-access-pfrsf\") pod \"ceilometer-0\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " pod="openstack/ceilometer-0" Dec 10 09:53:33 crc kubenswrapper[4715]: I1210 09:53:33.835576 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:34 crc kubenswrapper[4715]: W1210 09:53:34.324104 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda50402ab_ad6d_4031_9beb_a9fb8d4e74e8.slice/crio-f296021e115ff47c403fbad5f8c2689928a4d4fec32b001347bb688823c98777 WatchSource:0}: Error finding container f296021e115ff47c403fbad5f8c2689928a4d4fec32b001347bb688823c98777: Status 404 returned error can't find the container with id f296021e115ff47c403fbad5f8c2689928a4d4fec32b001347bb688823c98777 Dec 10 09:53:34 crc kubenswrapper[4715]: I1210 09:53:34.326102 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:34 crc kubenswrapper[4715]: I1210 09:53:34.384866 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8","Type":"ContainerStarted","Data":"f296021e115ff47c403fbad5f8c2689928a4d4fec32b001347bb688823c98777"} Dec 10 09:53:35 crc kubenswrapper[4715]: I1210 09:53:35.246245 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:36 crc kubenswrapper[4715]: I1210 09:53:36.405526 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8","Type":"ContainerStarted","Data":"0f8a10c33f846c4bd71af1cbce35ccb7044f454ebedf7e99af70046553b193b7"} Dec 10 09:53:36 crc kubenswrapper[4715]: I1210 09:53:36.406433 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8","Type":"ContainerStarted","Data":"647de7d363f4dc25b2d40a44426ba924af30358f085b9cd913f8a1d36705e365"} Dec 10 09:53:37 crc kubenswrapper[4715]: I1210 09:53:37.418011 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8","Type":"ContainerStarted","Data":"dd7b198a33e8ade36cbfbeb666210960365fc86d41fb5d2573207fe997060283"} Dec 10 09:53:38 crc kubenswrapper[4715]: I1210 09:53:38.089856 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 09:53:38 crc kubenswrapper[4715]: I1210 09:53:38.090161 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 09:53:38 crc kubenswrapper[4715]: I1210 09:53:38.126153 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 09:53:38 crc kubenswrapper[4715]: I1210 09:53:38.137035 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 09:53:38 crc kubenswrapper[4715]: I1210 09:53:38.429448 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 09:53:38 crc kubenswrapper[4715]: I1210 09:53:38.430695 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 09:53:38 crc kubenswrapper[4715]: I1210 09:53:38.905106 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:38 crc kubenswrapper[4715]: I1210 09:53:38.905173 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:38 crc kubenswrapper[4715]: I1210 09:53:38.946694 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:38 crc kubenswrapper[4715]: I1210 09:53:38.968088 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:39 crc kubenswrapper[4715]: I1210 09:53:39.441746 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:39 crc kubenswrapper[4715]: I1210 09:53:39.441804 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.465937 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8","Type":"ContainerStarted","Data":"3adc11b3a41771cfb3c1666b2a4616613801c28b443ed71c0a38607add864ed0"} Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.468782 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.466713 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="proxy-httpd" containerID="cri-o://3adc11b3a41771cfb3c1666b2a4616613801c28b443ed71c0a38607add864ed0" gracePeriod=30 Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.466739 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="sg-core" containerID="cri-o://dd7b198a33e8ade36cbfbeb666210960365fc86d41fb5d2573207fe997060283" gracePeriod=30 Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.466749 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="ceilometer-notification-agent" containerID="cri-o://0f8a10c33f846c4bd71af1cbce35ccb7044f454ebedf7e99af70046553b193b7" gracePeriod=30 Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.466792 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.469177 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.466081 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="ceilometer-central-agent" containerID="cri-o://647de7d363f4dc25b2d40a44426ba924af30358f085b9cd913f8a1d36705e365" gracePeriod=30 Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.566142 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.610439 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.498908398 podStartE2EDuration="7.610419729s" podCreationTimestamp="2025-12-10 09:53:33 +0000 UTC" firstStartedPulling="2025-12-10 09:53:34.327475957 +0000 UTC m=+1177.071022208" lastFinishedPulling="2025-12-10 09:53:39.438987248 +0000 UTC m=+1182.182533539" observedRunningTime="2025-12-10 09:53:40.49958985 +0000 UTC m=+1183.243136101" watchObservedRunningTime="2025-12-10 09:53:40.610419729 +0000 UTC m=+1183.353965980" Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.776378 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-s6r2g"] Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.777612 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-s6r2g" Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.799580 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-s6r2g"] Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.907217 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90981cd9-8647-4ee0-abc4-fdf840af114f-operator-scripts\") pod \"nova-api-db-create-s6r2g\" (UID: \"90981cd9-8647-4ee0-abc4-fdf840af114f\") " pod="openstack/nova-api-db-create-s6r2g" Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.907385 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn5xs\" (UniqueName: \"kubernetes.io/projected/90981cd9-8647-4ee0-abc4-fdf840af114f-kube-api-access-wn5xs\") pod \"nova-api-db-create-s6r2g\" (UID: \"90981cd9-8647-4ee0-abc4-fdf840af114f\") " pod="openstack/nova-api-db-create-s6r2g" Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.921063 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-l8nbx"] Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.949578 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-l8nbx" Dec 10 09:53:40 crc kubenswrapper[4715]: I1210 09:53:40.952591 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-l8nbx"] Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.009501 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqq2w\" (UniqueName: \"kubernetes.io/projected/8a223f35-4f28-4207-b900-4738035524c1-kube-api-access-gqq2w\") pod \"nova-cell0-db-create-l8nbx\" (UID: \"8a223f35-4f28-4207-b900-4738035524c1\") " pod="openstack/nova-cell0-db-create-l8nbx" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.009800 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn5xs\" (UniqueName: \"kubernetes.io/projected/90981cd9-8647-4ee0-abc4-fdf840af114f-kube-api-access-wn5xs\") pod \"nova-api-db-create-s6r2g\" (UID: \"90981cd9-8647-4ee0-abc4-fdf840af114f\") " pod="openstack/nova-api-db-create-s6r2g" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.009913 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a223f35-4f28-4207-b900-4738035524c1-operator-scripts\") pod \"nova-cell0-db-create-l8nbx\" (UID: \"8a223f35-4f28-4207-b900-4738035524c1\") " pod="openstack/nova-cell0-db-create-l8nbx" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.010046 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90981cd9-8647-4ee0-abc4-fdf840af114f-operator-scripts\") pod \"nova-api-db-create-s6r2g\" (UID: \"90981cd9-8647-4ee0-abc4-fdf840af114f\") " pod="openstack/nova-api-db-create-s6r2g" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.010877 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90981cd9-8647-4ee0-abc4-fdf840af114f-operator-scripts\") pod \"nova-api-db-create-s6r2g\" (UID: \"90981cd9-8647-4ee0-abc4-fdf840af114f\") " pod="openstack/nova-api-db-create-s6r2g" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.026983 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-dvpzw"] Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.028739 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dvpzw" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.041356 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn5xs\" (UniqueName: \"kubernetes.io/projected/90981cd9-8647-4ee0-abc4-fdf840af114f-kube-api-access-wn5xs\") pod \"nova-api-db-create-s6r2g\" (UID: \"90981cd9-8647-4ee0-abc4-fdf840af114f\") " pod="openstack/nova-api-db-create-s6r2g" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.054670 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dvpzw"] Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.089043 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-ccce-account-create-update-8pckm"] Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.093769 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ccce-account-create-update-8pckm" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.103130 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-s6r2g" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.103742 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.113823 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqq2w\" (UniqueName: \"kubernetes.io/projected/8a223f35-4f28-4207-b900-4738035524c1-kube-api-access-gqq2w\") pod \"nova-cell0-db-create-l8nbx\" (UID: \"8a223f35-4f28-4207-b900-4738035524c1\") " pod="openstack/nova-cell0-db-create-l8nbx" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.113872 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/284a9082-8b05-470c-93f3-1623b2a569dc-operator-scripts\") pod \"nova-cell1-db-create-dvpzw\" (UID: \"284a9082-8b05-470c-93f3-1623b2a569dc\") " pod="openstack/nova-cell1-db-create-dvpzw" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.122304 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a223f35-4f28-4207-b900-4738035524c1-operator-scripts\") pod \"nova-cell0-db-create-l8nbx\" (UID: \"8a223f35-4f28-4207-b900-4738035524c1\") " pod="openstack/nova-cell0-db-create-l8nbx" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.122575 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx66n\" (UniqueName: \"kubernetes.io/projected/284a9082-8b05-470c-93f3-1623b2a569dc-kube-api-access-jx66n\") pod \"nova-cell1-db-create-dvpzw\" (UID: \"284a9082-8b05-470c-93f3-1623b2a569dc\") " pod="openstack/nova-cell1-db-create-dvpzw" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.123716 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a223f35-4f28-4207-b900-4738035524c1-operator-scripts\") pod \"nova-cell0-db-create-l8nbx\" (UID: \"8a223f35-4f28-4207-b900-4738035524c1\") " pod="openstack/nova-cell0-db-create-l8nbx" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.123759 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-ccce-account-create-update-8pckm"] Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.156206 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqq2w\" (UniqueName: \"kubernetes.io/projected/8a223f35-4f28-4207-b900-4738035524c1-kube-api-access-gqq2w\") pod \"nova-cell0-db-create-l8nbx\" (UID: \"8a223f35-4f28-4207-b900-4738035524c1\") " pod="openstack/nova-cell0-db-create-l8nbx" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.213353 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-b909-account-create-update-f7lvf"] Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.214903 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b909-account-create-update-f7lvf" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.224855 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b909-account-create-update-f7lvf"] Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.225160 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.226899 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-operator-scripts\") pod \"nova-api-ccce-account-create-update-8pckm\" (UID: \"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52\") " pod="openstack/nova-api-ccce-account-create-update-8pckm" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.227064 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/284a9082-8b05-470c-93f3-1623b2a569dc-operator-scripts\") pod \"nova-cell1-db-create-dvpzw\" (UID: \"284a9082-8b05-470c-93f3-1623b2a569dc\") " pod="openstack/nova-cell1-db-create-dvpzw" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.227194 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nxwr\" (UniqueName: \"kubernetes.io/projected/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-kube-api-access-4nxwr\") pod \"nova-api-ccce-account-create-update-8pckm\" (UID: \"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52\") " pod="openstack/nova-api-ccce-account-create-update-8pckm" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.227348 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx66n\" (UniqueName: \"kubernetes.io/projected/284a9082-8b05-470c-93f3-1623b2a569dc-kube-api-access-jx66n\") pod \"nova-cell1-db-create-dvpzw\" (UID: \"284a9082-8b05-470c-93f3-1623b2a569dc\") " pod="openstack/nova-cell1-db-create-dvpzw" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.234178 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/284a9082-8b05-470c-93f3-1623b2a569dc-operator-scripts\") pod \"nova-cell1-db-create-dvpzw\" (UID: \"284a9082-8b05-470c-93f3-1623b2a569dc\") " pod="openstack/nova-cell1-db-create-dvpzw" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.260273 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx66n\" (UniqueName: \"kubernetes.io/projected/284a9082-8b05-470c-93f3-1623b2a569dc-kube-api-access-jx66n\") pod \"nova-cell1-db-create-dvpzw\" (UID: \"284a9082-8b05-470c-93f3-1623b2a569dc\") " pod="openstack/nova-cell1-db-create-dvpzw" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.289175 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-l8nbx" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.331803 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-operator-scripts\") pod \"nova-api-ccce-account-create-update-8pckm\" (UID: \"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52\") " pod="openstack/nova-api-ccce-account-create-update-8pckm" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.331888 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-operator-scripts\") pod \"nova-cell0-b909-account-create-update-f7lvf\" (UID: \"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f\") " pod="openstack/nova-cell0-b909-account-create-update-f7lvf" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.331963 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nxwr\" (UniqueName: \"kubernetes.io/projected/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-kube-api-access-4nxwr\") pod \"nova-api-ccce-account-create-update-8pckm\" (UID: \"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52\") " pod="openstack/nova-api-ccce-account-create-update-8pckm" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.332007 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9gsp\" (UniqueName: \"kubernetes.io/projected/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-kube-api-access-n9gsp\") pod \"nova-cell0-b909-account-create-update-f7lvf\" (UID: \"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f\") " pod="openstack/nova-cell0-b909-account-create-update-f7lvf" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.333266 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-operator-scripts\") pod \"nova-api-ccce-account-create-update-8pckm\" (UID: \"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52\") " pod="openstack/nova-api-ccce-account-create-update-8pckm" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.362945 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nxwr\" (UniqueName: \"kubernetes.io/projected/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-kube-api-access-4nxwr\") pod \"nova-api-ccce-account-create-update-8pckm\" (UID: \"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52\") " pod="openstack/nova-api-ccce-account-create-update-8pckm" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.415393 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-1365-account-create-update-h5zsg"] Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.417194 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1365-account-create-update-h5zsg" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.420996 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.427779 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dvpzw" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.437270 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1365-account-create-update-h5zsg"] Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.442249 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.456923 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-operator-scripts\") pod \"nova-cell0-b909-account-create-update-f7lvf\" (UID: \"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f\") " pod="openstack/nova-cell0-b909-account-create-update-f7lvf" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.457002 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9gsp\" (UniqueName: \"kubernetes.io/projected/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-kube-api-access-n9gsp\") pod \"nova-cell0-b909-account-create-update-f7lvf\" (UID: \"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f\") " pod="openstack/nova-cell0-b909-account-create-update-f7lvf" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.457893 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-operator-scripts\") pod \"nova-cell0-b909-account-create-update-f7lvf\" (UID: \"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f\") " pod="openstack/nova-cell0-b909-account-create-update-f7lvf" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.489003 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9gsp\" (UniqueName: \"kubernetes.io/projected/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-kube-api-access-n9gsp\") pod \"nova-cell0-b909-account-create-update-f7lvf\" (UID: \"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f\") " pod="openstack/nova-cell0-b909-account-create-update-f7lvf" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.507589 4715 generic.go:334] "Generic (PLEG): container finished" podID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerID="3adc11b3a41771cfb3c1666b2a4616613801c28b443ed71c0a38607add864ed0" exitCode=0 Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.507619 4715 generic.go:334] "Generic (PLEG): container finished" podID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerID="dd7b198a33e8ade36cbfbeb666210960365fc86d41fb5d2573207fe997060283" exitCode=2 Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.507627 4715 generic.go:334] "Generic (PLEG): container finished" podID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerID="0f8a10c33f846c4bd71af1cbce35ccb7044f454ebedf7e99af70046553b193b7" exitCode=0 Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.509107 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8","Type":"ContainerDied","Data":"3adc11b3a41771cfb3c1666b2a4616613801c28b443ed71c0a38607add864ed0"} Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.509144 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8","Type":"ContainerDied","Data":"dd7b198a33e8ade36cbfbeb666210960365fc86d41fb5d2573207fe997060283"} Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.509156 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8","Type":"ContainerDied","Data":"0f8a10c33f846c4bd71af1cbce35ccb7044f454ebedf7e99af70046553b193b7"} Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.509216 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.509228 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.570205 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8sgl\" (UniqueName: \"kubernetes.io/projected/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-kube-api-access-h8sgl\") pod \"nova-cell1-1365-account-create-update-h5zsg\" (UID: \"e97c28af-aa74-4d5e-9c58-6bf64e02c4df\") " pod="openstack/nova-cell1-1365-account-create-update-h5zsg" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.570394 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-operator-scripts\") pod \"nova-cell1-1365-account-create-update-h5zsg\" (UID: \"e97c28af-aa74-4d5e-9c58-6bf64e02c4df\") " pod="openstack/nova-cell1-1365-account-create-update-h5zsg" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.581306 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ccce-account-create-update-8pckm" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.587391 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b909-account-create-update-f7lvf" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.671849 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-operator-scripts\") pod \"nova-cell1-1365-account-create-update-h5zsg\" (UID: \"e97c28af-aa74-4d5e-9c58-6bf64e02c4df\") " pod="openstack/nova-cell1-1365-account-create-update-h5zsg" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.672098 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8sgl\" (UniqueName: \"kubernetes.io/projected/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-kube-api-access-h8sgl\") pod \"nova-cell1-1365-account-create-update-h5zsg\" (UID: \"e97c28af-aa74-4d5e-9c58-6bf64e02c4df\") " pod="openstack/nova-cell1-1365-account-create-update-h5zsg" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.677427 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-operator-scripts\") pod \"nova-cell1-1365-account-create-update-h5zsg\" (UID: \"e97c28af-aa74-4d5e-9c58-6bf64e02c4df\") " pod="openstack/nova-cell1-1365-account-create-update-h5zsg" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.702861 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8sgl\" (UniqueName: \"kubernetes.io/projected/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-kube-api-access-h8sgl\") pod \"nova-cell1-1365-account-create-update-h5zsg\" (UID: \"e97c28af-aa74-4d5e-9c58-6bf64e02c4df\") " pod="openstack/nova-cell1-1365-account-create-update-h5zsg" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.764308 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1365-account-create-update-h5zsg" Dec 10 09:53:41 crc kubenswrapper[4715]: I1210 09:53:41.774476 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-s6r2g"] Dec 10 09:53:41 crc kubenswrapper[4715]: W1210 09:53:41.848377 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90981cd9_8647_4ee0_abc4_fdf840af114f.slice/crio-be6e4a49f1961f0946ae2bfb33fb43ebf44f5d5d0ab667cc3541d8d2612dc7da WatchSource:0}: Error finding container be6e4a49f1961f0946ae2bfb33fb43ebf44f5d5d0ab667cc3541d8d2612dc7da: Status 404 returned error can't find the container with id be6e4a49f1961f0946ae2bfb33fb43ebf44f5d5d0ab667cc3541d8d2612dc7da Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.033644 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-l8nbx"] Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.267558 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dvpzw"] Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.297288 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b909-account-create-update-f7lvf"] Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.425089 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-ccce-account-create-update-8pckm"] Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.506971 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1365-account-create-update-h5zsg"] Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.561738 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b909-account-create-update-f7lvf" event={"ID":"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f","Type":"ContainerStarted","Data":"a3cf9caa360da5474654adc83d50b5cb0bbeead59950db4727663996c5638870"} Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.566607 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dvpzw" event={"ID":"284a9082-8b05-470c-93f3-1623b2a569dc","Type":"ContainerStarted","Data":"78038a3157086e42207fc0bf00fbc6e74570431fdca7963cacf975ac23572479"} Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.566661 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dvpzw" event={"ID":"284a9082-8b05-470c-93f3-1623b2a569dc","Type":"ContainerStarted","Data":"726a0bfcfbbb0bf2e17bd6ba842570e0356e5621a15588f7a816df184fe8dd05"} Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.592217 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ccce-account-create-update-8pckm" event={"ID":"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52","Type":"ContainerStarted","Data":"1898ba76461579766f0c09516c5fb063a22930427130492abb2cdfea91f49943"} Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.595687 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-s6r2g" event={"ID":"90981cd9-8647-4ee0-abc4-fdf840af114f","Type":"ContainerStarted","Data":"61ff90633b7fde6006561463ae4b159e0ac49361f01f389e9161c22d5fd7982f"} Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.595733 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-s6r2g" event={"ID":"90981cd9-8647-4ee0-abc4-fdf840af114f","Type":"ContainerStarted","Data":"be6e4a49f1961f0946ae2bfb33fb43ebf44f5d5d0ab667cc3541d8d2612dc7da"} Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.625391 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-l8nbx" event={"ID":"8a223f35-4f28-4207-b900-4738035524c1","Type":"ContainerStarted","Data":"a9df0962aca3eacbd049ad154eb47491a466190b45070a5f3c3b8d462739b884"} Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.625435 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-l8nbx" event={"ID":"8a223f35-4f28-4207-b900-4738035524c1","Type":"ContainerStarted","Data":"da63df4afc0a15b2726742723c6af41178dae729e6317c06c10ab063fce9a6f6"} Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.646385 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-dvpzw" podStartSLOduration=2.646357932 podStartE2EDuration="2.646357932s" podCreationTimestamp="2025-12-10 09:53:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:53:42.620719011 +0000 UTC m=+1185.364265262" watchObservedRunningTime="2025-12-10 09:53:42.646357932 +0000 UTC m=+1185.389904183" Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.722620 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-s6r2g" podStartSLOduration=2.722605356 podStartE2EDuration="2.722605356s" podCreationTimestamp="2025-12-10 09:53:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:53:42.646527057 +0000 UTC m=+1185.390073328" watchObservedRunningTime="2025-12-10 09:53:42.722605356 +0000 UTC m=+1185.466151607" Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.756682 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-l8nbx" podStartSLOduration=2.756661046 podStartE2EDuration="2.756661046s" podCreationTimestamp="2025-12-10 09:53:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:53:42.686096475 +0000 UTC m=+1185.429642726" watchObservedRunningTime="2025-12-10 09:53:42.756661046 +0000 UTC m=+1185.500207297" Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.910369 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.910709 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 09:53:42 crc kubenswrapper[4715]: I1210 09:53:42.911216 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.630859 4715 generic.go:334] "Generic (PLEG): container finished" podID="284a9082-8b05-470c-93f3-1623b2a569dc" containerID="78038a3157086e42207fc0bf00fbc6e74570431fdca7963cacf975ac23572479" exitCode=0 Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.630955 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dvpzw" event={"ID":"284a9082-8b05-470c-93f3-1623b2a569dc","Type":"ContainerDied","Data":"78038a3157086e42207fc0bf00fbc6e74570431fdca7963cacf975ac23572479"} Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.634898 4715 generic.go:334] "Generic (PLEG): container finished" podID="b5bc47a9-7f16-4f49-b2e1-904dcdae3b52" containerID="5714ec46d17f9b4828df90f75ed2d929ebaca9ce49cd494e8ab89ee2a5e49a88" exitCode=0 Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.634984 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ccce-account-create-update-8pckm" event={"ID":"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52","Type":"ContainerDied","Data":"5714ec46d17f9b4828df90f75ed2d929ebaca9ce49cd494e8ab89ee2a5e49a88"} Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.636776 4715 generic.go:334] "Generic (PLEG): container finished" podID="90981cd9-8647-4ee0-abc4-fdf840af114f" containerID="61ff90633b7fde6006561463ae4b159e0ac49361f01f389e9161c22d5fd7982f" exitCode=0 Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.636849 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-s6r2g" event={"ID":"90981cd9-8647-4ee0-abc4-fdf840af114f","Type":"ContainerDied","Data":"61ff90633b7fde6006561463ae4b159e0ac49361f01f389e9161c22d5fd7982f"} Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.638467 4715 generic.go:334] "Generic (PLEG): container finished" podID="8a223f35-4f28-4207-b900-4738035524c1" containerID="a9df0962aca3eacbd049ad154eb47491a466190b45070a5f3c3b8d462739b884" exitCode=0 Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.638527 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-l8nbx" event={"ID":"8a223f35-4f28-4207-b900-4738035524c1","Type":"ContainerDied","Data":"a9df0962aca3eacbd049ad154eb47491a466190b45070a5f3c3b8d462739b884"} Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.640483 4715 generic.go:334] "Generic (PLEG): container finished" podID="e97c28af-aa74-4d5e-9c58-6bf64e02c4df" containerID="078d33857fc129aa312a0999a6a5b26d216ea2d9d7f87e691900288c52c38f2a" exitCode=0 Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.640538 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1365-account-create-update-h5zsg" event={"ID":"e97c28af-aa74-4d5e-9c58-6bf64e02c4df","Type":"ContainerDied","Data":"078d33857fc129aa312a0999a6a5b26d216ea2d9d7f87e691900288c52c38f2a"} Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.640557 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1365-account-create-update-h5zsg" event={"ID":"e97c28af-aa74-4d5e-9c58-6bf64e02c4df","Type":"ContainerStarted","Data":"1f2ae1b4304969df2f548c3412452f474fe15e1481dc2ed3013b0bad00eedbd1"} Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.641893 4715 generic.go:334] "Generic (PLEG): container finished" podID="d7b6c95a-d6f0-4753-bcaa-08cecbe2289f" containerID="d35a3305f328c3e6e0e741d01efbfbd21dc4a0fda9f38e2536971940ee98434d" exitCode=0 Dec 10 09:53:43 crc kubenswrapper[4715]: I1210 09:53:43.641979 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b909-account-create-update-f7lvf" event={"ID":"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f","Type":"ContainerDied","Data":"d35a3305f328c3e6e0e741d01efbfbd21dc4a0fda9f38e2536971940ee98434d"} Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.133302 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b909-account-create-update-f7lvf" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.172604 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-operator-scripts\") pod \"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f\" (UID: \"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.172711 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9gsp\" (UniqueName: \"kubernetes.io/projected/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-kube-api-access-n9gsp\") pod \"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f\" (UID: \"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.175675 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7b6c95a-d6f0-4753-bcaa-08cecbe2289f" (UID: "d7b6c95a-d6f0-4753-bcaa-08cecbe2289f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.190320 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-kube-api-access-n9gsp" (OuterVolumeSpecName: "kube-api-access-n9gsp") pod "d7b6c95a-d6f0-4753-bcaa-08cecbe2289f" (UID: "d7b6c95a-d6f0-4753-bcaa-08cecbe2289f"). InnerVolumeSpecName "kube-api-access-n9gsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.279078 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.279493 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9gsp\" (UniqueName: \"kubernetes.io/projected/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f-kube-api-access-n9gsp\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.378937 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dvpzw" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.402934 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ccce-account-create-update-8pckm" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.412940 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-s6r2g" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.423150 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1365-account-create-update-h5zsg" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.438259 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-l8nbx" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.482310 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nxwr\" (UniqueName: \"kubernetes.io/projected/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-kube-api-access-4nxwr\") pod \"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52\" (UID: \"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.482383 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-operator-scripts\") pod \"e97c28af-aa74-4d5e-9c58-6bf64e02c4df\" (UID: \"e97c28af-aa74-4d5e-9c58-6bf64e02c4df\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.482415 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90981cd9-8647-4ee0-abc4-fdf840af114f-operator-scripts\") pod \"90981cd9-8647-4ee0-abc4-fdf840af114f\" (UID: \"90981cd9-8647-4ee0-abc4-fdf840af114f\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.482509 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn5xs\" (UniqueName: \"kubernetes.io/projected/90981cd9-8647-4ee0-abc4-fdf840af114f-kube-api-access-wn5xs\") pod \"90981cd9-8647-4ee0-abc4-fdf840af114f\" (UID: \"90981cd9-8647-4ee0-abc4-fdf840af114f\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.482537 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8sgl\" (UniqueName: \"kubernetes.io/projected/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-kube-api-access-h8sgl\") pod \"e97c28af-aa74-4d5e-9c58-6bf64e02c4df\" (UID: \"e97c28af-aa74-4d5e-9c58-6bf64e02c4df\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.482574 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-operator-scripts\") pod \"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52\" (UID: \"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.482686 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/284a9082-8b05-470c-93f3-1623b2a569dc-operator-scripts\") pod \"284a9082-8b05-470c-93f3-1623b2a569dc\" (UID: \"284a9082-8b05-470c-93f3-1623b2a569dc\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.482744 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx66n\" (UniqueName: \"kubernetes.io/projected/284a9082-8b05-470c-93f3-1623b2a569dc-kube-api-access-jx66n\") pod \"284a9082-8b05-470c-93f3-1623b2a569dc\" (UID: \"284a9082-8b05-470c-93f3-1623b2a569dc\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.482810 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a223f35-4f28-4207-b900-4738035524c1-operator-scripts\") pod \"8a223f35-4f28-4207-b900-4738035524c1\" (UID: \"8a223f35-4f28-4207-b900-4738035524c1\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.482881 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqq2w\" (UniqueName: \"kubernetes.io/projected/8a223f35-4f28-4207-b900-4738035524c1-kube-api-access-gqq2w\") pod \"8a223f35-4f28-4207-b900-4738035524c1\" (UID: \"8a223f35-4f28-4207-b900-4738035524c1\") " Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.484607 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/284a9082-8b05-470c-93f3-1623b2a569dc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "284a9082-8b05-470c-93f3-1623b2a569dc" (UID: "284a9082-8b05-470c-93f3-1623b2a569dc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.484970 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5bc47a9-7f16-4f49-b2e1-904dcdae3b52" (UID: "b5bc47a9-7f16-4f49-b2e1-904dcdae3b52"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.485301 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90981cd9-8647-4ee0-abc4-fdf840af114f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "90981cd9-8647-4ee0-abc4-fdf840af114f" (UID: "90981cd9-8647-4ee0-abc4-fdf840af114f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.485632 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e97c28af-aa74-4d5e-9c58-6bf64e02c4df" (UID: "e97c28af-aa74-4d5e-9c58-6bf64e02c4df"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.485961 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a223f35-4f28-4207-b900-4738035524c1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a223f35-4f28-4207-b900-4738035524c1" (UID: "8a223f35-4f28-4207-b900-4738035524c1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.487613 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a223f35-4f28-4207-b900-4738035524c1-kube-api-access-gqq2w" (OuterVolumeSpecName: "kube-api-access-gqq2w") pod "8a223f35-4f28-4207-b900-4738035524c1" (UID: "8a223f35-4f28-4207-b900-4738035524c1"). InnerVolumeSpecName "kube-api-access-gqq2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.490318 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-kube-api-access-h8sgl" (OuterVolumeSpecName: "kube-api-access-h8sgl") pod "e97c28af-aa74-4d5e-9c58-6bf64e02c4df" (UID: "e97c28af-aa74-4d5e-9c58-6bf64e02c4df"). InnerVolumeSpecName "kube-api-access-h8sgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.490596 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90981cd9-8647-4ee0-abc4-fdf840af114f-kube-api-access-wn5xs" (OuterVolumeSpecName: "kube-api-access-wn5xs") pod "90981cd9-8647-4ee0-abc4-fdf840af114f" (UID: "90981cd9-8647-4ee0-abc4-fdf840af114f"). InnerVolumeSpecName "kube-api-access-wn5xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.490610 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-kube-api-access-4nxwr" (OuterVolumeSpecName: "kube-api-access-4nxwr") pod "b5bc47a9-7f16-4f49-b2e1-904dcdae3b52" (UID: "b5bc47a9-7f16-4f49-b2e1-904dcdae3b52"). InnerVolumeSpecName "kube-api-access-4nxwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.504099 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/284a9082-8b05-470c-93f3-1623b2a569dc-kube-api-access-jx66n" (OuterVolumeSpecName: "kube-api-access-jx66n") pod "284a9082-8b05-470c-93f3-1623b2a569dc" (UID: "284a9082-8b05-470c-93f3-1623b2a569dc"). InnerVolumeSpecName "kube-api-access-jx66n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.585458 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqq2w\" (UniqueName: \"kubernetes.io/projected/8a223f35-4f28-4207-b900-4738035524c1-kube-api-access-gqq2w\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.585496 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nxwr\" (UniqueName: \"kubernetes.io/projected/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-kube-api-access-4nxwr\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.585508 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.585518 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90981cd9-8647-4ee0-abc4-fdf840af114f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.585527 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn5xs\" (UniqueName: \"kubernetes.io/projected/90981cd9-8647-4ee0-abc4-fdf840af114f-kube-api-access-wn5xs\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.585535 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8sgl\" (UniqueName: \"kubernetes.io/projected/e97c28af-aa74-4d5e-9c58-6bf64e02c4df-kube-api-access-h8sgl\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.585544 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.585554 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/284a9082-8b05-470c-93f3-1623b2a569dc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.585562 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx66n\" (UniqueName: \"kubernetes.io/projected/284a9082-8b05-470c-93f3-1623b2a569dc-kube-api-access-jx66n\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.585569 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a223f35-4f28-4207-b900-4738035524c1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.663029 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b909-account-create-update-f7lvf" event={"ID":"d7b6c95a-d6f0-4753-bcaa-08cecbe2289f","Type":"ContainerDied","Data":"a3cf9caa360da5474654adc83d50b5cb0bbeead59950db4727663996c5638870"} Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.663081 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3cf9caa360da5474654adc83d50b5cb0bbeead59950db4727663996c5638870" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.663144 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b909-account-create-update-f7lvf" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.669648 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dvpzw" event={"ID":"284a9082-8b05-470c-93f3-1623b2a569dc","Type":"ContainerDied","Data":"726a0bfcfbbb0bf2e17bd6ba842570e0356e5621a15588f7a816df184fe8dd05"} Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.669693 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="726a0bfcfbbb0bf2e17bd6ba842570e0356e5621a15588f7a816df184fe8dd05" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.669691 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dvpzw" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.671400 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ccce-account-create-update-8pckm" event={"ID":"b5bc47a9-7f16-4f49-b2e1-904dcdae3b52","Type":"ContainerDied","Data":"1898ba76461579766f0c09516c5fb063a22930427130492abb2cdfea91f49943"} Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.671439 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1898ba76461579766f0c09516c5fb063a22930427130492abb2cdfea91f49943" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.671851 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ccce-account-create-update-8pckm" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.673137 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-s6r2g" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.673151 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-s6r2g" event={"ID":"90981cd9-8647-4ee0-abc4-fdf840af114f","Type":"ContainerDied","Data":"be6e4a49f1961f0946ae2bfb33fb43ebf44f5d5d0ab667cc3541d8d2612dc7da"} Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.673186 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be6e4a49f1961f0946ae2bfb33fb43ebf44f5d5d0ab667cc3541d8d2612dc7da" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.674616 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-l8nbx" event={"ID":"8a223f35-4f28-4207-b900-4738035524c1","Type":"ContainerDied","Data":"da63df4afc0a15b2726742723c6af41178dae729e6317c06c10ab063fce9a6f6"} Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.674638 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da63df4afc0a15b2726742723c6af41178dae729e6317c06c10ab063fce9a6f6" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.674664 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-l8nbx" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.677768 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1365-account-create-update-h5zsg" event={"ID":"e97c28af-aa74-4d5e-9c58-6bf64e02c4df","Type":"ContainerDied","Data":"1f2ae1b4304969df2f548c3412452f474fe15e1481dc2ed3013b0bad00eedbd1"} Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.678700 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f2ae1b4304969df2f548c3412452f474fe15e1481dc2ed3013b0bad00eedbd1" Dec 10 09:53:45 crc kubenswrapper[4715]: I1210 09:53:45.677836 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1365-account-create-update-h5zsg" Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.698182 4715 generic.go:334] "Generic (PLEG): container finished" podID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerID="647de7d363f4dc25b2d40a44426ba924af30358f085b9cd913f8a1d36705e365" exitCode=0 Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.698276 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8","Type":"ContainerDied","Data":"647de7d363f4dc25b2d40a44426ba924af30358f085b9cd913f8a1d36705e365"} Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.880870 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.909032 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-scripts\") pod \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.909184 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-log-httpd\") pod \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.909337 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfrsf\" (UniqueName: \"kubernetes.io/projected/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-kube-api-access-pfrsf\") pod \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.909392 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-config-data\") pod \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.909435 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-run-httpd\") pod \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.909498 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-sg-core-conf-yaml\") pod \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.909519 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-combined-ca-bundle\") pod \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\" (UID: \"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8\") " Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.909732 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" (UID: "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.910107 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.910326 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" (UID: "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.918803 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-kube-api-access-pfrsf" (OuterVolumeSpecName: "kube-api-access-pfrsf") pod "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" (UID: "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8"). InnerVolumeSpecName "kube-api-access-pfrsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.931118 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-scripts" (OuterVolumeSpecName: "scripts") pod "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" (UID: "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:46 crc kubenswrapper[4715]: I1210 09:53:46.939080 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" (UID: "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.005150 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" (UID: "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.011286 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.011496 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.011576 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.011643 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.011701 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfrsf\" (UniqueName: \"kubernetes.io/projected/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-kube-api-access-pfrsf\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.025146 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-config-data" (OuterVolumeSpecName: "config-data") pod "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" (UID: "a50402ab-ad6d-4031-9beb-a9fb8d4e74e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.113092 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.710801 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a50402ab-ad6d-4031-9beb-a9fb8d4e74e8","Type":"ContainerDied","Data":"f296021e115ff47c403fbad5f8c2689928a4d4fec32b001347bb688823c98777"} Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.710873 4715 scope.go:117] "RemoveContainer" containerID="3adc11b3a41771cfb3c1666b2a4616613801c28b443ed71c0a38607add864ed0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.710876 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.715390 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.715481 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.737308 4715 scope.go:117] "RemoveContainer" containerID="dd7b198a33e8ade36cbfbeb666210960365fc86d41fb5d2573207fe997060283" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.740738 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.755654 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.764586 4715 scope.go:117] "RemoveContainer" containerID="0f8a10c33f846c4bd71af1cbce35ccb7044f454ebedf7e99af70046553b193b7" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.776094 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:47 crc kubenswrapper[4715]: E1210 09:53:47.777217 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a223f35-4f28-4207-b900-4738035524c1" containerName="mariadb-database-create" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777244 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a223f35-4f28-4207-b900-4738035524c1" containerName="mariadb-database-create" Dec 10 09:53:47 crc kubenswrapper[4715]: E1210 09:53:47.777267 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="sg-core" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777274 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="sg-core" Dec 10 09:53:47 crc kubenswrapper[4715]: E1210 09:53:47.777297 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="ceilometer-central-agent" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777304 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="ceilometer-central-agent" Dec 10 09:53:47 crc kubenswrapper[4715]: E1210 09:53:47.777316 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b6c95a-d6f0-4753-bcaa-08cecbe2289f" containerName="mariadb-account-create-update" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777323 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b6c95a-d6f0-4753-bcaa-08cecbe2289f" containerName="mariadb-account-create-update" Dec 10 09:53:47 crc kubenswrapper[4715]: E1210 09:53:47.777337 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="proxy-httpd" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777343 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="proxy-httpd" Dec 10 09:53:47 crc kubenswrapper[4715]: E1210 09:53:47.777353 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90981cd9-8647-4ee0-abc4-fdf840af114f" containerName="mariadb-database-create" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777359 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="90981cd9-8647-4ee0-abc4-fdf840af114f" containerName="mariadb-database-create" Dec 10 09:53:47 crc kubenswrapper[4715]: E1210 09:53:47.777385 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e97c28af-aa74-4d5e-9c58-6bf64e02c4df" containerName="mariadb-account-create-update" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777391 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e97c28af-aa74-4d5e-9c58-6bf64e02c4df" containerName="mariadb-account-create-update" Dec 10 09:53:47 crc kubenswrapper[4715]: E1210 09:53:47.777413 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5bc47a9-7f16-4f49-b2e1-904dcdae3b52" containerName="mariadb-account-create-update" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777419 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5bc47a9-7f16-4f49-b2e1-904dcdae3b52" containerName="mariadb-account-create-update" Dec 10 09:53:47 crc kubenswrapper[4715]: E1210 09:53:47.777437 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284a9082-8b05-470c-93f3-1623b2a569dc" containerName="mariadb-database-create" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777443 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="284a9082-8b05-470c-93f3-1623b2a569dc" containerName="mariadb-database-create" Dec 10 09:53:47 crc kubenswrapper[4715]: E1210 09:53:47.777460 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="ceilometer-notification-agent" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777471 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="ceilometer-notification-agent" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777829 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5bc47a9-7f16-4f49-b2e1-904dcdae3b52" containerName="mariadb-account-create-update" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777853 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="sg-core" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777858 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b6c95a-d6f0-4753-bcaa-08cecbe2289f" containerName="mariadb-account-create-update" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777878 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a223f35-4f28-4207-b900-4738035524c1" containerName="mariadb-database-create" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777896 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="284a9082-8b05-470c-93f3-1623b2a569dc" containerName="mariadb-database-create" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777935 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="ceilometer-central-agent" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777946 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="proxy-httpd" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777965 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="90981cd9-8647-4ee0-abc4-fdf840af114f" containerName="mariadb-database-create" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777984 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" containerName="ceilometer-notification-agent" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.777995 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e97c28af-aa74-4d5e-9c58-6bf64e02c4df" containerName="mariadb-account-create-update" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.783516 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.791428 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.791999 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.792809 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.810025 4715 scope.go:117] "RemoveContainer" containerID="647de7d363f4dc25b2d40a44426ba924af30358f085b9cd913f8a1d36705e365" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.826069 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.826135 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-config-data\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.826213 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.826459 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-scripts\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.826566 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-run-httpd\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.826827 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shj88\" (UniqueName: \"kubernetes.io/projected/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-kube-api-access-shj88\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.826894 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-log-httpd\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.929285 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shj88\" (UniqueName: \"kubernetes.io/projected/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-kube-api-access-shj88\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.929355 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-log-httpd\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.929469 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.929521 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-config-data\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.929597 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.929624 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-scripts\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.929694 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-run-httpd\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.930067 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-log-httpd\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.930376 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-run-httpd\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.934652 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.935148 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-config-data\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.941829 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-scripts\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.948217 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:47 crc kubenswrapper[4715]: I1210 09:53:47.949008 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shj88\" (UniqueName: \"kubernetes.io/projected/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-kube-api-access-shj88\") pod \"ceilometer-0\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " pod="openstack/ceilometer-0" Dec 10 09:53:48 crc kubenswrapper[4715]: I1210 09:53:48.111586 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:53:48 crc kubenswrapper[4715]: I1210 09:53:48.590665 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:53:48 crc kubenswrapper[4715]: I1210 09:53:48.722508 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"770f9f0f-b5a0-4f3d-a585-91ed12aa874e","Type":"ContainerStarted","Data":"49b2f153e561155b0d5a78cea27d1e84b2ca31e3d410c672f0ddb42defb874ff"} Dec 10 09:53:49 crc kubenswrapper[4715]: I1210 09:53:49.616676 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a50402ab-ad6d-4031-9beb-a9fb8d4e74e8" path="/var/lib/kubelet/pods/a50402ab-ad6d-4031-9beb-a9fb8d4e74e8/volumes" Dec 10 09:53:49 crc kubenswrapper[4715]: I1210 09:53:49.733590 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"770f9f0f-b5a0-4f3d-a585-91ed12aa874e","Type":"ContainerStarted","Data":"83d45601b0b9a2b50f6f032644d5a95641a14f5c9103d638a58d4edb179674eb"} Dec 10 09:53:50 crc kubenswrapper[4715]: I1210 09:53:50.743588 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"770f9f0f-b5a0-4f3d-a585-91ed12aa874e","Type":"ContainerStarted","Data":"fba9262b625db5aaf4e7f990bc5320905f4e4f9343dc14c54ea7f9c1f1f670ae"} Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.443189 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kkzgm"] Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.444787 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.446738 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.446894 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.447017 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2ztt9" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.456119 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kkzgm"] Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.512309 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh9dg\" (UniqueName: \"kubernetes.io/projected/8c7977c2-7055-41e5-8d3a-ad1e652186bc-kube-api-access-jh9dg\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.512360 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-scripts\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.512437 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.512474 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-config-data\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.613548 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh9dg\" (UniqueName: \"kubernetes.io/projected/8c7977c2-7055-41e5-8d3a-ad1e652186bc-kube-api-access-jh9dg\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.613588 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-scripts\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.613656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.613682 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-config-data\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.618244 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-scripts\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.620260 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.635627 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh9dg\" (UniqueName: \"kubernetes.io/projected/8c7977c2-7055-41e5-8d3a-ad1e652186bc-kube-api-access-jh9dg\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.637005 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-config-data\") pod \"nova-cell0-conductor-db-sync-kkzgm\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.756872 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"770f9f0f-b5a0-4f3d-a585-91ed12aa874e","Type":"ContainerStarted","Data":"093fc62679b5d45882383a16e1918725d935a37e09282cdbfecb2af62c10e7fd"} Dec 10 09:53:51 crc kubenswrapper[4715]: I1210 09:53:51.766536 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:53:52 crc kubenswrapper[4715]: I1210 09:53:52.281967 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kkzgm"] Dec 10 09:53:52 crc kubenswrapper[4715]: I1210 09:53:52.768991 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"770f9f0f-b5a0-4f3d-a585-91ed12aa874e","Type":"ContainerStarted","Data":"d6302cf81087d77c068ea48b3cfd604814cda0c8abe9e8d0fe4a3c5a19a09b0a"} Dec 10 09:53:52 crc kubenswrapper[4715]: I1210 09:53:52.769391 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 09:53:52 crc kubenswrapper[4715]: I1210 09:53:52.773760 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kkzgm" event={"ID":"8c7977c2-7055-41e5-8d3a-ad1e652186bc","Type":"ContainerStarted","Data":"b2275ec8ac21ce63532bba06742de7359015a79cd51909ee919cad2615f4db8b"} Dec 10 09:53:52 crc kubenswrapper[4715]: I1210 09:53:52.799243 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.995936734 podStartE2EDuration="5.799221855s" podCreationTimestamp="2025-12-10 09:53:47 +0000 UTC" firstStartedPulling="2025-12-10 09:53:48.602292724 +0000 UTC m=+1191.345838975" lastFinishedPulling="2025-12-10 09:53:52.405577845 +0000 UTC m=+1195.149124096" observedRunningTime="2025-12-10 09:53:52.790525387 +0000 UTC m=+1195.534071638" watchObservedRunningTime="2025-12-10 09:53:52.799221855 +0000 UTC m=+1195.542768106" Dec 10 09:54:02 crc kubenswrapper[4715]: I1210 09:54:02.899267 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kkzgm" event={"ID":"8c7977c2-7055-41e5-8d3a-ad1e652186bc","Type":"ContainerStarted","Data":"01e548c196e50c07af9e0cff406b4e663addfa3a7e025b7b064262d8ccac4cae"} Dec 10 09:54:02 crc kubenswrapper[4715]: I1210 09:54:02.927070 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-kkzgm" podStartSLOduration=2.166016021 podStartE2EDuration="11.927050404s" podCreationTimestamp="2025-12-10 09:53:51 +0000 UTC" firstStartedPulling="2025-12-10 09:53:52.287656253 +0000 UTC m=+1195.031202514" lastFinishedPulling="2025-12-10 09:54:02.048690646 +0000 UTC m=+1204.792236897" observedRunningTime="2025-12-10 09:54:02.916304327 +0000 UTC m=+1205.659850588" watchObservedRunningTime="2025-12-10 09:54:02.927050404 +0000 UTC m=+1205.670596655" Dec 10 09:54:15 crc kubenswrapper[4715]: I1210 09:54:15.152081 4715 generic.go:334] "Generic (PLEG): container finished" podID="8c7977c2-7055-41e5-8d3a-ad1e652186bc" containerID="01e548c196e50c07af9e0cff406b4e663addfa3a7e025b7b064262d8ccac4cae" exitCode=0 Dec 10 09:54:15 crc kubenswrapper[4715]: I1210 09:54:15.152713 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kkzgm" event={"ID":"8c7977c2-7055-41e5-8d3a-ad1e652186bc","Type":"ContainerDied","Data":"01e548c196e50c07af9e0cff406b4e663addfa3a7e025b7b064262d8ccac4cae"} Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.565169 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.570463 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh9dg\" (UniqueName: \"kubernetes.io/projected/8c7977c2-7055-41e5-8d3a-ad1e652186bc-kube-api-access-jh9dg\") pod \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.570602 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-combined-ca-bundle\") pod \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.570748 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-config-data\") pod \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.570789 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-scripts\") pod \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\" (UID: \"8c7977c2-7055-41e5-8d3a-ad1e652186bc\") " Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.576720 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-scripts" (OuterVolumeSpecName: "scripts") pod "8c7977c2-7055-41e5-8d3a-ad1e652186bc" (UID: "8c7977c2-7055-41e5-8d3a-ad1e652186bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.577375 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c7977c2-7055-41e5-8d3a-ad1e652186bc-kube-api-access-jh9dg" (OuterVolumeSpecName: "kube-api-access-jh9dg") pod "8c7977c2-7055-41e5-8d3a-ad1e652186bc" (UID: "8c7977c2-7055-41e5-8d3a-ad1e652186bc"). InnerVolumeSpecName "kube-api-access-jh9dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.603856 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-config-data" (OuterVolumeSpecName: "config-data") pod "8c7977c2-7055-41e5-8d3a-ad1e652186bc" (UID: "8c7977c2-7055-41e5-8d3a-ad1e652186bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.604696 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c7977c2-7055-41e5-8d3a-ad1e652186bc" (UID: "8c7977c2-7055-41e5-8d3a-ad1e652186bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.673293 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.673320 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.673332 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7977c2-7055-41e5-8d3a-ad1e652186bc-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:16 crc kubenswrapper[4715]: I1210 09:54:16.673344 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh9dg\" (UniqueName: \"kubernetes.io/projected/8c7977c2-7055-41e5-8d3a-ad1e652186bc-kube-api-access-jh9dg\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.175155 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kkzgm" event={"ID":"8c7977c2-7055-41e5-8d3a-ad1e652186bc","Type":"ContainerDied","Data":"b2275ec8ac21ce63532bba06742de7359015a79cd51909ee919cad2615f4db8b"} Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.175205 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2275ec8ac21ce63532bba06742de7359015a79cd51909ee919cad2615f4db8b" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.175269 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kkzgm" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.278553 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 09:54:17 crc kubenswrapper[4715]: E1210 09:54:17.279087 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7977c2-7055-41e5-8d3a-ad1e652186bc" containerName="nova-cell0-conductor-db-sync" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.279113 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7977c2-7055-41e5-8d3a-ad1e652186bc" containerName="nova-cell0-conductor-db-sync" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.279336 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c7977c2-7055-41e5-8d3a-ad1e652186bc" containerName="nova-cell0-conductor-db-sync" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.280162 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.284079 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2ztt9" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.285386 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.292849 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 09:54:17 crc kubenswrapper[4715]: E1210 09:54:17.376442 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c7977c2_7055_41e5_8d3a_ad1e652186bc.slice\": RecentStats: unable to find data in memory cache]" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.383759 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da\") " pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.383836 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngx58\" (UniqueName: \"kubernetes.io/projected/6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da-kube-api-access-ngx58\") pod \"nova-cell0-conductor-0\" (UID: \"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da\") " pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.384431 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da\") " pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.487157 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da\") " pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.487215 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngx58\" (UniqueName: \"kubernetes.io/projected/6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da-kube-api-access-ngx58\") pod \"nova-cell0-conductor-0\" (UID: \"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da\") " pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.487324 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da\") " pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.493793 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da\") " pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.495195 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da\") " pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.504465 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngx58\" (UniqueName: \"kubernetes.io/projected/6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da-kube-api-access-ngx58\") pod \"nova-cell0-conductor-0\" (UID: \"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da\") " pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.605809 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.714113 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.714593 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.714676 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.715685 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"13f540c3510be87c668f6f3667fb9d329e2ff72443e0f54c1a4f373c716a3669"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 09:54:17 crc kubenswrapper[4715]: I1210 09:54:17.715888 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://13f540c3510be87c668f6f3667fb9d329e2ff72443e0f54c1a4f373c716a3669" gracePeriod=600 Dec 10 09:54:18 crc kubenswrapper[4715]: I1210 09:54:18.074061 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 09:54:18 crc kubenswrapper[4715]: W1210 09:54:18.077771 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f8bdb76_8ee4_43e6_b1c6_68f4ef3814da.slice/crio-058e0696cc32e885dbbba0cd0c8f56ee840d72af73cd024a9541eb2fa05b7144 WatchSource:0}: Error finding container 058e0696cc32e885dbbba0cd0c8f56ee840d72af73cd024a9541eb2fa05b7144: Status 404 returned error can't find the container with id 058e0696cc32e885dbbba0cd0c8f56ee840d72af73cd024a9541eb2fa05b7144 Dec 10 09:54:18 crc kubenswrapper[4715]: I1210 09:54:18.119132 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 09:54:18 crc kubenswrapper[4715]: I1210 09:54:18.201709 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da","Type":"ContainerStarted","Data":"058e0696cc32e885dbbba0cd0c8f56ee840d72af73cd024a9541eb2fa05b7144"} Dec 10 09:54:18 crc kubenswrapper[4715]: I1210 09:54:18.205183 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="13f540c3510be87c668f6f3667fb9d329e2ff72443e0f54c1a4f373c716a3669" exitCode=0 Dec 10 09:54:18 crc kubenswrapper[4715]: I1210 09:54:18.205276 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"13f540c3510be87c668f6f3667fb9d329e2ff72443e0f54c1a4f373c716a3669"} Dec 10 09:54:18 crc kubenswrapper[4715]: I1210 09:54:18.205351 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"43f0315f5b714d86cf23ed6f6c468531cbdb76f130137362d46c5a9fc1556442"} Dec 10 09:54:18 crc kubenswrapper[4715]: I1210 09:54:18.205388 4715 scope.go:117] "RemoveContainer" containerID="92c7512a624555c7386239654aadd7a7479838a440573c30b4fffcacf96a04bf" Dec 10 09:54:19 crc kubenswrapper[4715]: I1210 09:54:19.216820 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da","Type":"ContainerStarted","Data":"5150b526569303fd4e3402985563a40cc9afed34c5b36596b9ce840ff4650318"} Dec 10 09:54:19 crc kubenswrapper[4715]: I1210 09:54:19.218480 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:19 crc kubenswrapper[4715]: I1210 09:54:19.236821 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.236794196 podStartE2EDuration="2.236794196s" podCreationTimestamp="2025-12-10 09:54:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:54:19.235085877 +0000 UTC m=+1221.978632148" watchObservedRunningTime="2025-12-10 09:54:19.236794196 +0000 UTC m=+1221.980340447" Dec 10 09:54:22 crc kubenswrapper[4715]: I1210 09:54:22.028173 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 09:54:22 crc kubenswrapper[4715]: I1210 09:54:22.029328 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="66bffaf0-4efe-435f-8092-a14c323cb97d" containerName="kube-state-metrics" containerID="cri-o://c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a" gracePeriod=30 Dec 10 09:54:22 crc kubenswrapper[4715]: I1210 09:54:22.631734 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 09:54:22 crc kubenswrapper[4715]: I1210 09:54:22.798948 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q55hp\" (UniqueName: \"kubernetes.io/projected/66bffaf0-4efe-435f-8092-a14c323cb97d-kube-api-access-q55hp\") pod \"66bffaf0-4efe-435f-8092-a14c323cb97d\" (UID: \"66bffaf0-4efe-435f-8092-a14c323cb97d\") " Dec 10 09:54:22 crc kubenswrapper[4715]: I1210 09:54:22.807966 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66bffaf0-4efe-435f-8092-a14c323cb97d-kube-api-access-q55hp" (OuterVolumeSpecName: "kube-api-access-q55hp") pod "66bffaf0-4efe-435f-8092-a14c323cb97d" (UID: "66bffaf0-4efe-435f-8092-a14c323cb97d"). InnerVolumeSpecName "kube-api-access-q55hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:54:22 crc kubenswrapper[4715]: I1210 09:54:22.903116 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q55hp\" (UniqueName: \"kubernetes.io/projected/66bffaf0-4efe-435f-8092-a14c323cb97d-kube-api-access-q55hp\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.260871 4715 generic.go:334] "Generic (PLEG): container finished" podID="66bffaf0-4efe-435f-8092-a14c323cb97d" containerID="c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a" exitCode=2 Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.260924 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66bffaf0-4efe-435f-8092-a14c323cb97d","Type":"ContainerDied","Data":"c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a"} Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.260938 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.260959 4715 scope.go:117] "RemoveContainer" containerID="c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.260949 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"66bffaf0-4efe-435f-8092-a14c323cb97d","Type":"ContainerDied","Data":"0aec39697e8f2909b71d014aa11aa278abb71385d92caf441026b4467852c635"} Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.293297 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.294323 4715 scope.go:117] "RemoveContainer" containerID="c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a" Dec 10 09:54:23 crc kubenswrapper[4715]: E1210 09:54:23.294819 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a\": container with ID starting with c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a not found: ID does not exist" containerID="c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.294868 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a"} err="failed to get container status \"c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a\": rpc error: code = NotFound desc = could not find container \"c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a\": container with ID starting with c40918edc21c7c959c826687f557b2afb40e9790448ea53e691234eed050894a not found: ID does not exist" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.302452 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.328287 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 09:54:23 crc kubenswrapper[4715]: E1210 09:54:23.328852 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66bffaf0-4efe-435f-8092-a14c323cb97d" containerName="kube-state-metrics" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.328876 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="66bffaf0-4efe-435f-8092-a14c323cb97d" containerName="kube-state-metrics" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.329182 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="66bffaf0-4efe-435f-8092-a14c323cb97d" containerName="kube-state-metrics" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.330003 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.330087 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.336180 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.337013 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.412563 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf760357-b816-4b56-b56d-5a3637a7b2b0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.413233 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/cf760357-b816-4b56-b56d-5a3637a7b2b0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.413381 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ngb8\" (UniqueName: \"kubernetes.io/projected/cf760357-b816-4b56-b56d-5a3637a7b2b0-kube-api-access-5ngb8\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.413626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf760357-b816-4b56-b56d-5a3637a7b2b0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.515412 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/cf760357-b816-4b56-b56d-5a3637a7b2b0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.515743 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ngb8\" (UniqueName: \"kubernetes.io/projected/cf760357-b816-4b56-b56d-5a3637a7b2b0-kube-api-access-5ngb8\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.515836 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf760357-b816-4b56-b56d-5a3637a7b2b0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.516061 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf760357-b816-4b56-b56d-5a3637a7b2b0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.520587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf760357-b816-4b56-b56d-5a3637a7b2b0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.521091 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/cf760357-b816-4b56-b56d-5a3637a7b2b0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.539620 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf760357-b816-4b56-b56d-5a3637a7b2b0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.548706 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ngb8\" (UniqueName: \"kubernetes.io/projected/cf760357-b816-4b56-b56d-5a3637a7b2b0-kube-api-access-5ngb8\") pod \"kube-state-metrics-0\" (UID: \"cf760357-b816-4b56-b56d-5a3637a7b2b0\") " pod="openstack/kube-state-metrics-0" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.616125 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66bffaf0-4efe-435f-8092-a14c323cb97d" path="/var/lib/kubelet/pods/66bffaf0-4efe-435f-8092-a14c323cb97d/volumes" Dec 10 09:54:23 crc kubenswrapper[4715]: I1210 09:54:23.663621 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 09:54:24 crc kubenswrapper[4715]: I1210 09:54:24.079768 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 09:54:24 crc kubenswrapper[4715]: I1210 09:54:24.088747 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 09:54:24 crc kubenswrapper[4715]: I1210 09:54:24.212109 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:54:24 crc kubenswrapper[4715]: I1210 09:54:24.212441 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="ceilometer-central-agent" containerID="cri-o://83d45601b0b9a2b50f6f032644d5a95641a14f5c9103d638a58d4edb179674eb" gracePeriod=30 Dec 10 09:54:24 crc kubenswrapper[4715]: I1210 09:54:24.212537 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="proxy-httpd" containerID="cri-o://d6302cf81087d77c068ea48b3cfd604814cda0c8abe9e8d0fe4a3c5a19a09b0a" gracePeriod=30 Dec 10 09:54:24 crc kubenswrapper[4715]: I1210 09:54:24.212603 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="sg-core" containerID="cri-o://093fc62679b5d45882383a16e1918725d935a37e09282cdbfecb2af62c10e7fd" gracePeriod=30 Dec 10 09:54:24 crc kubenswrapper[4715]: I1210 09:54:24.212661 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="ceilometer-notification-agent" containerID="cri-o://fba9262b625db5aaf4e7f990bc5320905f4e4f9343dc14c54ea7f9c1f1f670ae" gracePeriod=30 Dec 10 09:54:24 crc kubenswrapper[4715]: I1210 09:54:24.271326 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cf760357-b816-4b56-b56d-5a3637a7b2b0","Type":"ContainerStarted","Data":"390cc13757e84f275f55cc06051010dc98d2fabaa550fda2923380efa5850361"} Dec 10 09:54:25 crc kubenswrapper[4715]: I1210 09:54:25.287414 4715 generic.go:334] "Generic (PLEG): container finished" podID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerID="d6302cf81087d77c068ea48b3cfd604814cda0c8abe9e8d0fe4a3c5a19a09b0a" exitCode=0 Dec 10 09:54:25 crc kubenswrapper[4715]: I1210 09:54:25.288393 4715 generic.go:334] "Generic (PLEG): container finished" podID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerID="093fc62679b5d45882383a16e1918725d935a37e09282cdbfecb2af62c10e7fd" exitCode=2 Dec 10 09:54:25 crc kubenswrapper[4715]: I1210 09:54:25.288410 4715 generic.go:334] "Generic (PLEG): container finished" podID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerID="83d45601b0b9a2b50f6f032644d5a95641a14f5c9103d638a58d4edb179674eb" exitCode=0 Dec 10 09:54:25 crc kubenswrapper[4715]: I1210 09:54:25.287468 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"770f9f0f-b5a0-4f3d-a585-91ed12aa874e","Type":"ContainerDied","Data":"d6302cf81087d77c068ea48b3cfd604814cda0c8abe9e8d0fe4a3c5a19a09b0a"} Dec 10 09:54:25 crc kubenswrapper[4715]: I1210 09:54:25.288461 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"770f9f0f-b5a0-4f3d-a585-91ed12aa874e","Type":"ContainerDied","Data":"093fc62679b5d45882383a16e1918725d935a37e09282cdbfecb2af62c10e7fd"} Dec 10 09:54:25 crc kubenswrapper[4715]: I1210 09:54:25.288481 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"770f9f0f-b5a0-4f3d-a585-91ed12aa874e","Type":"ContainerDied","Data":"83d45601b0b9a2b50f6f032644d5a95641a14f5c9103d638a58d4edb179674eb"} Dec 10 09:54:26 crc kubenswrapper[4715]: I1210 09:54:26.298037 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cf760357-b816-4b56-b56d-5a3637a7b2b0","Type":"ContainerStarted","Data":"e8f2fdd1ef5c1e85171acb0ee4162cdc42c38170c3c4dd660f332470da1f253c"} Dec 10 09:54:26 crc kubenswrapper[4715]: I1210 09:54:26.298195 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 10 09:54:26 crc kubenswrapper[4715]: I1210 09:54:26.326564 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.327728854 podStartE2EDuration="3.326538965s" podCreationTimestamp="2025-12-10 09:54:23 +0000 UTC" firstStartedPulling="2025-12-10 09:54:24.088503692 +0000 UTC m=+1226.832049943" lastFinishedPulling="2025-12-10 09:54:25.087313803 +0000 UTC m=+1227.830860054" observedRunningTime="2025-12-10 09:54:26.312229137 +0000 UTC m=+1229.055775398" watchObservedRunningTime="2025-12-10 09:54:26.326538965 +0000 UTC m=+1229.070085216" Dec 10 09:54:27 crc kubenswrapper[4715]: I1210 09:54:27.650846 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.275143 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-sbn8r"] Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.276836 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.279642 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.279655 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.287446 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-sbn8r"] Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.331752 4715 generic.go:334] "Generic (PLEG): container finished" podID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerID="fba9262b625db5aaf4e7f990bc5320905f4e4f9343dc14c54ea7f9c1f1f670ae" exitCode=0 Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.331806 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"770f9f0f-b5a0-4f3d-a585-91ed12aa874e","Type":"ContainerDied","Data":"fba9262b625db5aaf4e7f990bc5320905f4e4f9343dc14c54ea7f9c1f1f670ae"} Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.331837 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"770f9f0f-b5a0-4f3d-a585-91ed12aa874e","Type":"ContainerDied","Data":"49b2f153e561155b0d5a78cea27d1e84b2ca31e3d410c672f0ddb42defb874ff"} Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.331849 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49b2f153e561155b0d5a78cea27d1e84b2ca31e3d410c672f0ddb42defb874ff" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.338616 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412237 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-log-httpd\") pod \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412298 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shj88\" (UniqueName: \"kubernetes.io/projected/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-kube-api-access-shj88\") pod \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412362 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-run-httpd\") pod \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412381 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-sg-core-conf-yaml\") pod \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412426 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-config-data\") pod \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412448 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-scripts\") pod \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412565 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-combined-ca-bundle\") pod \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\" (UID: \"770f9f0f-b5a0-4f3d-a585-91ed12aa874e\") " Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412784 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wft5\" (UniqueName: \"kubernetes.io/projected/9936c060-cb6f-4c92-895b-c0381d18085e-kube-api-access-7wft5\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-scripts\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412889 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.412978 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-config-data\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.413210 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "770f9f0f-b5a0-4f3d-a585-91ed12aa874e" (UID: "770f9f0f-b5a0-4f3d-a585-91ed12aa874e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.413529 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "770f9f0f-b5a0-4f3d-a585-91ed12aa874e" (UID: "770f9f0f-b5a0-4f3d-a585-91ed12aa874e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.429397 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-kube-api-access-shj88" (OuterVolumeSpecName: "kube-api-access-shj88") pod "770f9f0f-b5a0-4f3d-a585-91ed12aa874e" (UID: "770f9f0f-b5a0-4f3d-a585-91ed12aa874e"). InnerVolumeSpecName "kube-api-access-shj88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.442271 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-scripts" (OuterVolumeSpecName: "scripts") pod "770f9f0f-b5a0-4f3d-a585-91ed12aa874e" (UID: "770f9f0f-b5a0-4f3d-a585-91ed12aa874e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.513392 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 09:54:28 crc kubenswrapper[4715]: E1210 09:54:28.513794 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="ceilometer-central-agent" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.513806 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="ceilometer-central-agent" Dec 10 09:54:28 crc kubenswrapper[4715]: E1210 09:54:28.513819 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="sg-core" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.513824 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="sg-core" Dec 10 09:54:28 crc kubenswrapper[4715]: E1210 09:54:28.513854 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="ceilometer-notification-agent" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.513860 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="ceilometer-notification-agent" Dec 10 09:54:28 crc kubenswrapper[4715]: E1210 09:54:28.513873 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="proxy-httpd" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.513878 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="proxy-httpd" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.514068 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="ceilometer-notification-agent" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.514089 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="proxy-httpd" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.514104 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="ceilometer-central-agent" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.514115 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" containerName="sg-core" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.515033 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.515089 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.515201 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-config-data\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.515268 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wft5\" (UniqueName: \"kubernetes.io/projected/9936c060-cb6f-4c92-895b-c0381d18085e-kube-api-access-7wft5\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.515351 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-scripts\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.515423 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.515442 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shj88\" (UniqueName: \"kubernetes.io/projected/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-kube-api-access-shj88\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.515455 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.515466 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.520461 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "770f9f0f-b5a0-4f3d-a585-91ed12aa874e" (UID: "770f9f0f-b5a0-4f3d-a585-91ed12aa874e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.541525 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.542885 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-scripts\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.563749 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-config-data\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.565569 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wft5\" (UniqueName: \"kubernetes.io/projected/9936c060-cb6f-4c92-895b-c0381d18085e-kube-api-access-7wft5\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.591499 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.599412 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-sbn8r\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.620592 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgs2f\" (UniqueName: \"kubernetes.io/projected/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-kube-api-access-lgs2f\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.620674 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-config-data\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.620811 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-logs\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.620830 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.620880 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.635011 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.644889 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.667537 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.680525 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.730364 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-logs\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.730486 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.730516 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww89l\" (UniqueName: \"kubernetes.io/projected/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-kube-api-access-ww89l\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.730578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-logs\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.730595 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.730615 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-config-data\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.730661 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgs2f\" (UniqueName: \"kubernetes.io/projected/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-kube-api-access-lgs2f\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.730803 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-config-data\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.732828 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.735629 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-logs\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.736842 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.755775 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-config-data\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.758299 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgs2f\" (UniqueName: \"kubernetes.io/projected/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-kube-api-access-lgs2f\") pod \"nova-api-0\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.808726 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "770f9f0f-b5a0-4f3d-a585-91ed12aa874e" (UID: "770f9f0f-b5a0-4f3d-a585-91ed12aa874e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.818643 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.820570 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.823568 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.835279 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.835334 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww89l\" (UniqueName: \"kubernetes.io/projected/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-kube-api-access-ww89l\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.835394 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-config-data\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.835651 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-logs\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.835712 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.838205 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-logs\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.845573 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-config-data\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.849893 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.853827 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-config-data" (OuterVolumeSpecName: "config-data") pod "770f9f0f-b5a0-4f3d-a585-91ed12aa874e" (UID: "770f9f0f-b5a0-4f3d-a585-91ed12aa874e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.863340 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.863981 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.864708 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.865874 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww89l\" (UniqueName: \"kubernetes.io/projected/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-kube-api-access-ww89l\") pod \"nova-metadata-0\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " pod="openstack/nova-metadata-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.867075 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.890481 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.906476 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.912816 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q6kwd"] Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.914689 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.951377 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-config-data\") pod \"nova-scheduler-0\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.951450 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsx5l\" (UniqueName: \"kubernetes.io/projected/cdba20b8-d67e-431f-bdcc-2763de1a1fea-kube-api-access-vsx5l\") pod \"nova-scheduler-0\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.951476 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.951568 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.951607 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.951645 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrblp\" (UniqueName: \"kubernetes.io/projected/aaad46b8-fed6-4ed4-9845-5ddfa954edad-kube-api-access-jrblp\") pod \"nova-cell1-novncproxy-0\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.951902 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/770f9f0f-b5a0-4f3d-a585-91ed12aa874e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:28 crc kubenswrapper[4715]: I1210 09:54:28.967138 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q6kwd"] Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.022351 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.053736 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-svc\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.053791 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46jkl\" (UniqueName: \"kubernetes.io/projected/7742e050-7183-4c20-9367-618075abadc8-kube-api-access-46jkl\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.053844 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-config\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.053901 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.053956 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.054009 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.054069 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrblp\" (UniqueName: \"kubernetes.io/projected/aaad46b8-fed6-4ed4-9845-5ddfa954edad-kube-api-access-jrblp\") pod \"nova-cell1-novncproxy-0\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.054154 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.054192 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-config-data\") pod \"nova-scheduler-0\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.054220 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.054274 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsx5l\" (UniqueName: \"kubernetes.io/projected/cdba20b8-d67e-431f-bdcc-2763de1a1fea-kube-api-access-vsx5l\") pod \"nova-scheduler-0\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.054302 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.061820 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.062398 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.071741 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-config-data\") pod \"nova-scheduler-0\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.072958 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.076473 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrblp\" (UniqueName: \"kubernetes.io/projected/aaad46b8-fed6-4ed4-9845-5ddfa954edad-kube-api-access-jrblp\") pod \"nova-cell1-novncproxy-0\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.083391 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsx5l\" (UniqueName: \"kubernetes.io/projected/cdba20b8-d67e-431f-bdcc-2763de1a1fea-kube-api-access-vsx5l\") pod \"nova-scheduler-0\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.142680 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.156747 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.156854 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-svc\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.156887 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46jkl\" (UniqueName: \"kubernetes.io/projected/7742e050-7183-4c20-9367-618075abadc8-kube-api-access-46jkl\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.156942 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-config\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.156989 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.157106 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.158250 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.158899 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-config\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.161167 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.161408 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-svc\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.162402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.196378 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46jkl\" (UniqueName: \"kubernetes.io/projected/7742e050-7183-4c20-9367-618075abadc8-kube-api-access-46jkl\") pod \"dnsmasq-dns-757b4f8459-q6kwd\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.196902 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.283219 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-sbn8r"] Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.300424 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.380301 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.380413 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-sbn8r" event={"ID":"9936c060-cb6f-4c92-895b-c0381d18085e","Type":"ContainerStarted","Data":"ff6ff74ed4d3f076df2566a9fc7570c254056f121f484ff05896fe57f4fc92ed"} Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.436988 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.452304 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.479809 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.483516 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.492568 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.496275 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.496544 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.545406 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.559968 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-thxxv"] Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.561433 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.565168 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.565404 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-config-data\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.565456 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-scripts\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.565475 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.565532 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.565574 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-log-httpd\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.565607 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzbvj\" (UniqueName: \"kubernetes.io/projected/3f30313c-cf6f-4ec5-a2da-70c9f7575918-kube-api-access-jzbvj\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.565695 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-run-httpd\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.565727 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.565836 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.577896 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-thxxv"] Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.638048 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="770f9f0f-b5a0-4f3d-a585-91ed12aa874e" path="/var/lib/kubelet/pods/770f9f0f-b5a0-4f3d-a585-91ed12aa874e/volumes" Dec 10 09:54:29 crc kubenswrapper[4715]: W1210 09:54:29.640497 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16c382ac_62d5_40b2_9fa3_fb6e57612f8a.slice/crio-8ba9d0335c62e3af04fadff58e6d64fd1b55689b296ee81f9e11ba10d332986d WatchSource:0}: Error finding container 8ba9d0335c62e3af04fadff58e6d64fd1b55689b296ee81f9e11ba10d332986d: Status 404 returned error can't find the container with id 8ba9d0335c62e3af04fadff58e6d64fd1b55689b296ee81f9e11ba10d332986d Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.640664 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.668957 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.672881 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-log-httpd\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.673074 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4rgm\" (UniqueName: \"kubernetes.io/projected/af79009e-1855-4f22-8397-b1c762175b72-kube-api-access-h4rgm\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.673224 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzbvj\" (UniqueName: \"kubernetes.io/projected/3f30313c-cf6f-4ec5-a2da-70c9f7575918-kube-api-access-jzbvj\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.673480 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-run-httpd\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.673648 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.673780 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.673980 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-config-data\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.674194 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-scripts\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.674321 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-config-data\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.674488 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-scripts\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.674609 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.674736 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-run-httpd\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.674493 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-log-httpd\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.677449 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.679566 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.680438 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-scripts\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.682219 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-config-data\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.685478 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.692249 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzbvj\" (UniqueName: \"kubernetes.io/projected/3f30313c-cf6f-4ec5-a2da-70c9f7575918-kube-api-access-jzbvj\") pod \"ceilometer-0\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: W1210 09:54:29.743620 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3ecfdf5_8dee_4ad8_bf3f_478c8d22ab8b.slice/crio-2454dcfae795241b505990830f84634ca26dba1ce47e4d9269ac4c94d82fccb7 WatchSource:0}: Error finding container 2454dcfae795241b505990830f84634ca26dba1ce47e4d9269ac4c94d82fccb7: Status 404 returned error can't find the container with id 2454dcfae795241b505990830f84634ca26dba1ce47e4d9269ac4c94d82fccb7 Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.759203 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.776361 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.776453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-config-data\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.776496 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-scripts\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.776586 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4rgm\" (UniqueName: \"kubernetes.io/projected/af79009e-1855-4f22-8397-b1c762175b72-kube-api-access-h4rgm\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.782152 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.784116 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-config-data\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.790412 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-scripts\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.799848 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4rgm\" (UniqueName: \"kubernetes.io/projected/af79009e-1855-4f22-8397-b1c762175b72-kube-api-access-h4rgm\") pod \"nova-cell1-conductor-db-sync-thxxv\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.823092 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.908842 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:29 crc kubenswrapper[4715]: I1210 09:54:29.946351 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.077984 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q6kwd"] Dec 10 09:54:30 crc kubenswrapper[4715]: W1210 09:54:30.098152 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdba20b8_d67e_431f_bdcc_2763de1a1fea.slice/crio-2fd32fdf2717004e6cc59df90c266d388cd34ec3be145964a20bf048df983f10 WatchSource:0}: Error finding container 2fd32fdf2717004e6cc59df90c266d388cd34ec3be145964a20bf048df983f10: Status 404 returned error can't find the container with id 2fd32fdf2717004e6cc59df90c266d388cd34ec3be145964a20bf048df983f10 Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.121516 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.333836 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.404474 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16c382ac-62d5-40b2-9fa3-fb6e57612f8a","Type":"ContainerStarted","Data":"8ba9d0335c62e3af04fadff58e6d64fd1b55689b296ee81f9e11ba10d332986d"} Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.408507 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" event={"ID":"7742e050-7183-4c20-9367-618075abadc8","Type":"ContainerStarted","Data":"ab98d930b1a679fce407e8a91fdd6197361eec9fc3f11fcbe8bd5b6ca3cd6b44"} Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.412309 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"aaad46b8-fed6-4ed4-9845-5ddfa954edad","Type":"ContainerStarted","Data":"409ea07dc02cfbffddc4bafb9b667fb5d3a7858b6adac1bb0b290a3f0b6d172a"} Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.415756 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f30313c-cf6f-4ec5-a2da-70c9f7575918","Type":"ContainerStarted","Data":"973035c580cd48cc8c345c4d7875e312558bcea9aacf89efe07cd671cf39ba92"} Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.421943 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cdba20b8-d67e-431f-bdcc-2763de1a1fea","Type":"ContainerStarted","Data":"2fd32fdf2717004e6cc59df90c266d388cd34ec3be145964a20bf048df983f10"} Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.436648 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-sbn8r" event={"ID":"9936c060-cb6f-4c92-895b-c0381d18085e","Type":"ContainerStarted","Data":"2dc1de321a39040ba4cb0ada4179d63413bbe3f996f2449d85212d2663e1099d"} Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.441157 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b","Type":"ContainerStarted","Data":"2454dcfae795241b505990830f84634ca26dba1ce47e4d9269ac4c94d82fccb7"} Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.472903 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-sbn8r" podStartSLOduration=2.472876695 podStartE2EDuration="2.472876695s" podCreationTimestamp="2025-12-10 09:54:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:54:30.460248675 +0000 UTC m=+1233.203794936" watchObservedRunningTime="2025-12-10 09:54:30.472876695 +0000 UTC m=+1233.216422946" Dec 10 09:54:30 crc kubenswrapper[4715]: I1210 09:54:30.552782 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-thxxv"] Dec 10 09:54:30 crc kubenswrapper[4715]: W1210 09:54:30.584626 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf79009e_1855_4f22_8397_b1c762175b72.slice/crio-99fef4ee72eacf461fd9ea40182279fa22948f51ba3ca250bab9600da3cbdfc2 WatchSource:0}: Error finding container 99fef4ee72eacf461fd9ea40182279fa22948f51ba3ca250bab9600da3cbdfc2: Status 404 returned error can't find the container with id 99fef4ee72eacf461fd9ea40182279fa22948f51ba3ca250bab9600da3cbdfc2 Dec 10 09:54:31 crc kubenswrapper[4715]: I1210 09:54:31.456182 4715 generic.go:334] "Generic (PLEG): container finished" podID="7742e050-7183-4c20-9367-618075abadc8" containerID="e5f46352a854686008a562fc489d85290d8a61903fa5abeb4653f35219f49bd2" exitCode=0 Dec 10 09:54:31 crc kubenswrapper[4715]: I1210 09:54:31.456375 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" event={"ID":"7742e050-7183-4c20-9367-618075abadc8","Type":"ContainerDied","Data":"e5f46352a854686008a562fc489d85290d8a61903fa5abeb4653f35219f49bd2"} Dec 10 09:54:31 crc kubenswrapper[4715]: I1210 09:54:31.463973 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-thxxv" event={"ID":"af79009e-1855-4f22-8397-b1c762175b72","Type":"ContainerStarted","Data":"6d9766176de8ce5757d3983f0f9746dce51ce47c6ecc7e1151cf20b35bc7c17a"} Dec 10 09:54:31 crc kubenswrapper[4715]: I1210 09:54:31.464022 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-thxxv" event={"ID":"af79009e-1855-4f22-8397-b1c762175b72","Type":"ContainerStarted","Data":"99fef4ee72eacf461fd9ea40182279fa22948f51ba3ca250bab9600da3cbdfc2"} Dec 10 09:54:31 crc kubenswrapper[4715]: I1210 09:54:31.505886 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-thxxv" podStartSLOduration=2.505846739 podStartE2EDuration="2.505846739s" podCreationTimestamp="2025-12-10 09:54:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:54:31.499497628 +0000 UTC m=+1234.243043889" watchObservedRunningTime="2025-12-10 09:54:31.505846739 +0000 UTC m=+1234.249392990" Dec 10 09:54:32 crc kubenswrapper[4715]: I1210 09:54:32.501265 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" event={"ID":"7742e050-7183-4c20-9367-618075abadc8","Type":"ContainerStarted","Data":"776f87f56130cb887230b8f13542b66aed7e25ffb3cb49306209bd93f96e9ca7"} Dec 10 09:54:32 crc kubenswrapper[4715]: I1210 09:54:32.501641 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:32 crc kubenswrapper[4715]: I1210 09:54:32.529792 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" podStartSLOduration=4.529769475 podStartE2EDuration="4.529769475s" podCreationTimestamp="2025-12-10 09:54:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:54:32.521728256 +0000 UTC m=+1235.265274507" watchObservedRunningTime="2025-12-10 09:54:32.529769475 +0000 UTC m=+1235.273315726" Dec 10 09:54:32 crc kubenswrapper[4715]: I1210 09:54:32.984767 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 09:54:32 crc kubenswrapper[4715]: I1210 09:54:32.997670 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:33 crc kubenswrapper[4715]: I1210 09:54:33.676552 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.566825 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16c382ac-62d5-40b2-9fa3-fb6e57612f8a","Type":"ContainerStarted","Data":"9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c"} Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.567414 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16c382ac-62d5-40b2-9fa3-fb6e57612f8a","Type":"ContainerStarted","Data":"d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a"} Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.569369 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"aaad46b8-fed6-4ed4-9845-5ddfa954edad","Type":"ContainerStarted","Data":"34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2"} Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.569470 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="aaad46b8-fed6-4ed4-9845-5ddfa954edad" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2" gracePeriod=30 Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.576535 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f30313c-cf6f-4ec5-a2da-70c9f7575918","Type":"ContainerStarted","Data":"c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c"} Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.579050 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cdba20b8-d67e-431f-bdcc-2763de1a1fea","Type":"ContainerStarted","Data":"a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23"} Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.586994 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b","Type":"ContainerStarted","Data":"a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd"} Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.587038 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b","Type":"ContainerStarted","Data":"e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4"} Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.587167 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" containerName="nova-metadata-log" containerID="cri-o://e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4" gracePeriod=30 Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.587250 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" containerName="nova-metadata-metadata" containerID="cri-o://a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd" gracePeriod=30 Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.600369 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.569415105 podStartE2EDuration="8.600352565s" podCreationTimestamp="2025-12-10 09:54:28 +0000 UTC" firstStartedPulling="2025-12-10 09:54:29.645587733 +0000 UTC m=+1232.389133984" lastFinishedPulling="2025-12-10 09:54:35.676525183 +0000 UTC m=+1238.420071444" observedRunningTime="2025-12-10 09:54:36.590823424 +0000 UTC m=+1239.334369675" watchObservedRunningTime="2025-12-10 09:54:36.600352565 +0000 UTC m=+1239.343898816" Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.647160 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.995718357 podStartE2EDuration="8.647123858s" podCreationTimestamp="2025-12-10 09:54:28 +0000 UTC" firstStartedPulling="2025-12-10 09:54:30.026986705 +0000 UTC m=+1232.770532956" lastFinishedPulling="2025-12-10 09:54:35.678392206 +0000 UTC m=+1238.421938457" observedRunningTime="2025-12-10 09:54:36.630314179 +0000 UTC m=+1239.373860430" watchObservedRunningTime="2025-12-10 09:54:36.647123858 +0000 UTC m=+1239.390670109" Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.667786 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.739271506 podStartE2EDuration="8.667754716s" podCreationTimestamp="2025-12-10 09:54:28 +0000 UTC" firstStartedPulling="2025-12-10 09:54:29.749547086 +0000 UTC m=+1232.493093327" lastFinishedPulling="2025-12-10 09:54:35.678030286 +0000 UTC m=+1238.421576537" observedRunningTime="2025-12-10 09:54:36.653538911 +0000 UTC m=+1239.397085162" watchObservedRunningTime="2025-12-10 09:54:36.667754716 +0000 UTC m=+1239.411300977" Dec 10 09:54:36 crc kubenswrapper[4715]: I1210 09:54:36.701022 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.148711049 podStartE2EDuration="8.700984424s" podCreationTimestamp="2025-12-10 09:54:28 +0000 UTC" firstStartedPulling="2025-12-10 09:54:30.126176112 +0000 UTC m=+1232.869722353" lastFinishedPulling="2025-12-10 09:54:35.678449477 +0000 UTC m=+1238.421995728" observedRunningTime="2025-12-10 09:54:36.680262273 +0000 UTC m=+1239.423808544" watchObservedRunningTime="2025-12-10 09:54:36.700984424 +0000 UTC m=+1239.444530675" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.476301 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.613907 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww89l\" (UniqueName: \"kubernetes.io/projected/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-kube-api-access-ww89l\") pod \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.614024 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-config-data\") pod \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.614147 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-logs\") pod \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.614179 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-combined-ca-bundle\") pod \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\" (UID: \"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b\") " Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.615830 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-logs" (OuterVolumeSpecName: "logs") pod "c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" (UID: "c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.635086 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-kube-api-access-ww89l" (OuterVolumeSpecName: "kube-api-access-ww89l") pod "c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" (UID: "c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b"). InnerVolumeSpecName "kube-api-access-ww89l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.647370 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f30313c-cf6f-4ec5-a2da-70c9f7575918","Type":"ContainerStarted","Data":"98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b"} Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.660739 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-config-data" (OuterVolumeSpecName: "config-data") pod "c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" (UID: "c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.662309 4715 generic.go:334] "Generic (PLEG): container finished" podID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" containerID="a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd" exitCode=0 Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.662352 4715 generic.go:334] "Generic (PLEG): container finished" podID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" containerID="e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4" exitCode=143 Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.662941 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b","Type":"ContainerDied","Data":"a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd"} Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.663073 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b","Type":"ContainerDied","Data":"e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4"} Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.663100 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b","Type":"ContainerDied","Data":"2454dcfae795241b505990830f84634ca26dba1ce47e4d9269ac4c94d82fccb7"} Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.663132 4715 scope.go:117] "RemoveContainer" containerID="a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.663519 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.672412 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" (UID: "c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.717764 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww89l\" (UniqueName: \"kubernetes.io/projected/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-kube-api-access-ww89l\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.717845 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.717862 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.717875 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:37 crc kubenswrapper[4715]: I1210 09:54:37.719705 4715 scope.go:117] "RemoveContainer" containerID="e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.004775 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.022429 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.035002 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:38 crc kubenswrapper[4715]: E1210 09:54:38.035802 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" containerName="nova-metadata-metadata" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.035825 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" containerName="nova-metadata-metadata" Dec 10 09:54:38 crc kubenswrapper[4715]: E1210 09:54:38.035872 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" containerName="nova-metadata-log" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.035881 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" containerName="nova-metadata-log" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.036202 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" containerName="nova-metadata-metadata" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.036246 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" containerName="nova-metadata-log" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.041463 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.044480 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.044864 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.072469 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.128515 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-config-data\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.128719 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l6mx\" (UniqueName: \"kubernetes.io/projected/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-kube-api-access-8l6mx\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.128778 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.128895 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-logs\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.128967 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.230593 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l6mx\" (UniqueName: \"kubernetes.io/projected/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-kube-api-access-8l6mx\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.230669 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.230770 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-logs\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.230825 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.230884 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-config-data\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.231380 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-logs\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.236197 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.236931 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-config-data\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.238080 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.264159 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l6mx\" (UniqueName: \"kubernetes.io/projected/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-kube-api-access-8l6mx\") pod \"nova-metadata-0\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.368043 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.558041 4715 scope.go:117] "RemoveContainer" containerID="a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd" Dec 10 09:54:38 crc kubenswrapper[4715]: E1210 09:54:38.562397 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd\": container with ID starting with a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd not found: ID does not exist" containerID="a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.562447 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd"} err="failed to get container status \"a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd\": rpc error: code = NotFound desc = could not find container \"a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd\": container with ID starting with a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd not found: ID does not exist" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.562478 4715 scope.go:117] "RemoveContainer" containerID="e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4" Dec 10 09:54:38 crc kubenswrapper[4715]: E1210 09:54:38.563139 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4\": container with ID starting with e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4 not found: ID does not exist" containerID="e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.563164 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4"} err="failed to get container status \"e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4\": rpc error: code = NotFound desc = could not find container \"e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4\": container with ID starting with e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4 not found: ID does not exist" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.563185 4715 scope.go:117] "RemoveContainer" containerID="a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.566044 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd"} err="failed to get container status \"a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd\": rpc error: code = NotFound desc = could not find container \"a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd\": container with ID starting with a29c3e8130b35f10ee7f216e40c91ba9965cc602799f667aec7514c4b893b9bd not found: ID does not exist" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.566113 4715 scope.go:117] "RemoveContainer" containerID="e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.566543 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4"} err="failed to get container status \"e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4\": rpc error: code = NotFound desc = could not find container \"e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4\": container with ID starting with e3dfa36a25cb7fa8ea50eed2f944ea7e528d4c3e34193d3e4b9eb8a906e8b7a4 not found: ID does not exist" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.911496 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.912056 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 09:54:38 crc kubenswrapper[4715]: I1210 09:54:38.990597 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.143525 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.143861 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.195177 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.198064 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.303645 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.391974 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qg7wp"] Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.396337 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" podUID="af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" containerName="dnsmasq-dns" containerID="cri-o://b0f842f4ca7b98a1bb87100e3351267e5cd609805b83ea442e3fad1d45914b7a" gracePeriod=10 Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.629146 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b" path="/var/lib/kubelet/pods/c3ecfdf5-8dee-4ad8-bf3f-478c8d22ab8b/volumes" Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.721200 4715 generic.go:334] "Generic (PLEG): container finished" podID="af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" containerID="b0f842f4ca7b98a1bb87100e3351267e5cd609805b83ea442e3fad1d45914b7a" exitCode=0 Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.721278 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" event={"ID":"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25","Type":"ContainerDied","Data":"b0f842f4ca7b98a1bb87100e3351267e5cd609805b83ea442e3fad1d45914b7a"} Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.725185 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f421ee4e-891c-4fff-b52a-3d0288e2b5c2","Type":"ContainerStarted","Data":"f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104"} Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.725260 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f421ee4e-891c-4fff-b52a-3d0288e2b5c2","Type":"ContainerStarted","Data":"a5d28f65bf6e5172395be3e177247b07cb123ef73242893c561b43c5e49da529"} Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.772711 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 09:54:39 crc kubenswrapper[4715]: I1210 09:54:39.871439 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" podUID="af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.162:5353: connect: connection refused" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.007776 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.008503 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.284684 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.299297 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjmfw\" (UniqueName: \"kubernetes.io/projected/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-kube-api-access-bjmfw\") pod \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.299852 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-nb\") pod \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.300120 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-swift-storage-0\") pod \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.300317 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-config\") pod \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.300410 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-sb\") pod \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.300536 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-svc\") pod \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\" (UID: \"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25\") " Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.316314 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-kube-api-access-bjmfw" (OuterVolumeSpecName: "kube-api-access-bjmfw") pod "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" (UID: "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25"). InnerVolumeSpecName "kube-api-access-bjmfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.378598 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-config" (OuterVolumeSpecName: "config") pod "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" (UID: "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.402541 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.402591 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjmfw\" (UniqueName: \"kubernetes.io/projected/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-kube-api-access-bjmfw\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.421433 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" (UID: "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.423875 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" (UID: "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.427026 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" (UID: "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.449395 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" (UID: "af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.504558 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.504996 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.505016 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.505031 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.736528 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" event={"ID":"af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25","Type":"ContainerDied","Data":"7dde7ea032f37ccc52e3c430b5de354a3b33123fbcf22765e9bed40c28d72cc8"} Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.736589 4715 scope.go:117] "RemoveContainer" containerID="b0f842f4ca7b98a1bb87100e3351267e5cd609805b83ea442e3fad1d45914b7a" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.736638 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-qg7wp" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.739577 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f30313c-cf6f-4ec5-a2da-70c9f7575918","Type":"ContainerStarted","Data":"1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491"} Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.746876 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f421ee4e-891c-4fff-b52a-3d0288e2b5c2","Type":"ContainerStarted","Data":"680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298"} Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.784136 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.784115941 podStartE2EDuration="2.784115941s" podCreationTimestamp="2025-12-10 09:54:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:54:40.78301553 +0000 UTC m=+1243.526561781" watchObservedRunningTime="2025-12-10 09:54:40.784115941 +0000 UTC m=+1243.527662192" Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.814150 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qg7wp"] Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.824476 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qg7wp"] Dec 10 09:54:40 crc kubenswrapper[4715]: I1210 09:54:40.852235 4715 scope.go:117] "RemoveContainer" containerID="4cf3d5f67480fb289d7598b0d540e92eaf74f47cb91e68dc19e767b6926f4794" Dec 10 09:54:41 crc kubenswrapper[4715]: I1210 09:54:41.620186 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" path="/var/lib/kubelet/pods/af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25/volumes" Dec 10 09:54:41 crc kubenswrapper[4715]: I1210 09:54:41.754285 4715 generic.go:334] "Generic (PLEG): container finished" podID="9936c060-cb6f-4c92-895b-c0381d18085e" containerID="2dc1de321a39040ba4cb0ada4179d63413bbe3f996f2449d85212d2663e1099d" exitCode=0 Dec 10 09:54:41 crc kubenswrapper[4715]: I1210 09:54:41.754520 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-sbn8r" event={"ID":"9936c060-cb6f-4c92-895b-c0381d18085e","Type":"ContainerDied","Data":"2dc1de321a39040ba4cb0ada4179d63413bbe3f996f2449d85212d2663e1099d"} Dec 10 09:54:42 crc kubenswrapper[4715]: I1210 09:54:42.766778 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f30313c-cf6f-4ec5-a2da-70c9f7575918","Type":"ContainerStarted","Data":"a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3"} Dec 10 09:54:42 crc kubenswrapper[4715]: I1210 09:54:42.767734 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 09:54:42 crc kubenswrapper[4715]: I1210 09:54:42.798298 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.02520768 podStartE2EDuration="13.798274415s" podCreationTimestamp="2025-12-10 09:54:29 +0000 UTC" firstStartedPulling="2025-12-10 09:54:30.353349988 +0000 UTC m=+1233.096896239" lastFinishedPulling="2025-12-10 09:54:42.126416733 +0000 UTC m=+1244.869962974" observedRunningTime="2025-12-10 09:54:42.786855059 +0000 UTC m=+1245.530401310" watchObservedRunningTime="2025-12-10 09:54:42.798274415 +0000 UTC m=+1245.541820666" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.223518 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.368484 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.369659 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.384463 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wft5\" (UniqueName: \"kubernetes.io/projected/9936c060-cb6f-4c92-895b-c0381d18085e-kube-api-access-7wft5\") pod \"9936c060-cb6f-4c92-895b-c0381d18085e\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.384512 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-scripts\") pod \"9936c060-cb6f-4c92-895b-c0381d18085e\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.384729 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-config-data\") pod \"9936c060-cb6f-4c92-895b-c0381d18085e\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.384867 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-combined-ca-bundle\") pod \"9936c060-cb6f-4c92-895b-c0381d18085e\" (UID: \"9936c060-cb6f-4c92-895b-c0381d18085e\") " Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.403798 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-scripts" (OuterVolumeSpecName: "scripts") pod "9936c060-cb6f-4c92-895b-c0381d18085e" (UID: "9936c060-cb6f-4c92-895b-c0381d18085e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.408940 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9936c060-cb6f-4c92-895b-c0381d18085e-kube-api-access-7wft5" (OuterVolumeSpecName: "kube-api-access-7wft5") pod "9936c060-cb6f-4c92-895b-c0381d18085e" (UID: "9936c060-cb6f-4c92-895b-c0381d18085e"). InnerVolumeSpecName "kube-api-access-7wft5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.414441 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9936c060-cb6f-4c92-895b-c0381d18085e" (UID: "9936c060-cb6f-4c92-895b-c0381d18085e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.418163 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-config-data" (OuterVolumeSpecName: "config-data") pod "9936c060-cb6f-4c92-895b-c0381d18085e" (UID: "9936c060-cb6f-4c92-895b-c0381d18085e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.487826 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wft5\" (UniqueName: \"kubernetes.io/projected/9936c060-cb6f-4c92-895b-c0381d18085e-kube-api-access-7wft5\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.487862 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.487871 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.487880 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9936c060-cb6f-4c92-895b-c0381d18085e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.778034 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-sbn8r" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.778464 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-sbn8r" event={"ID":"9936c060-cb6f-4c92-895b-c0381d18085e","Type":"ContainerDied","Data":"ff6ff74ed4d3f076df2566a9fc7570c254056f121f484ff05896fe57f4fc92ed"} Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.778483 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff6ff74ed4d3f076df2566a9fc7570c254056f121f484ff05896fe57f4fc92ed" Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.963485 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.963737 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerName="nova-api-log" containerID="cri-o://d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a" gracePeriod=30 Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.964248 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerName="nova-api-api" containerID="cri-o://9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c" gracePeriod=30 Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.978278 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:54:43 crc kubenswrapper[4715]: I1210 09:54:43.978489 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cdba20b8-d67e-431f-bdcc-2763de1a1fea" containerName="nova-scheduler-scheduler" containerID="cri-o://a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23" gracePeriod=30 Dec 10 09:54:44 crc kubenswrapper[4715]: I1210 09:54:44.009949 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:44 crc kubenswrapper[4715]: E1210 09:54:44.146305 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 09:54:44 crc kubenswrapper[4715]: E1210 09:54:44.148350 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 09:54:44 crc kubenswrapper[4715]: E1210 09:54:44.149567 4715 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 09:54:44 crc kubenswrapper[4715]: E1210 09:54:44.149614 4715 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cdba20b8-d67e-431f-bdcc-2763de1a1fea" containerName="nova-scheduler-scheduler" Dec 10 09:54:44 crc kubenswrapper[4715]: I1210 09:54:44.790845 4715 generic.go:334] "Generic (PLEG): container finished" podID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerID="d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a" exitCode=143 Dec 10 09:54:44 crc kubenswrapper[4715]: I1210 09:54:44.791081 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16c382ac-62d5-40b2-9fa3-fb6e57612f8a","Type":"ContainerDied","Data":"d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a"} Dec 10 09:54:44 crc kubenswrapper[4715]: I1210 09:54:44.793908 4715 generic.go:334] "Generic (PLEG): container finished" podID="af79009e-1855-4f22-8397-b1c762175b72" containerID="6d9766176de8ce5757d3983f0f9746dce51ce47c6ecc7e1151cf20b35bc7c17a" exitCode=0 Dec 10 09:54:44 crc kubenswrapper[4715]: I1210 09:54:44.793965 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-thxxv" event={"ID":"af79009e-1855-4f22-8397-b1c762175b72","Type":"ContainerDied","Data":"6d9766176de8ce5757d3983f0f9746dce51ce47c6ecc7e1151cf20b35bc7c17a"} Dec 10 09:54:44 crc kubenswrapper[4715]: I1210 09:54:44.794159 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" containerName="nova-metadata-log" containerID="cri-o://f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104" gracePeriod=30 Dec 10 09:54:44 crc kubenswrapper[4715]: I1210 09:54:44.794251 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" containerName="nova-metadata-metadata" containerID="cri-o://680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298" gracePeriod=30 Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.472486 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.573393 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-logs\") pod \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.573774 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-combined-ca-bundle\") pod \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.573759 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-logs" (OuterVolumeSpecName: "logs") pod "f421ee4e-891c-4fff-b52a-3d0288e2b5c2" (UID: "f421ee4e-891c-4fff-b52a-3d0288e2b5c2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.574096 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-config-data\") pod \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.574189 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l6mx\" (UniqueName: \"kubernetes.io/projected/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-kube-api-access-8l6mx\") pod \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.574296 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-nova-metadata-tls-certs\") pod \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\" (UID: \"f421ee4e-891c-4fff-b52a-3d0288e2b5c2\") " Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.574965 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.580749 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-kube-api-access-8l6mx" (OuterVolumeSpecName: "kube-api-access-8l6mx") pod "f421ee4e-891c-4fff-b52a-3d0288e2b5c2" (UID: "f421ee4e-891c-4fff-b52a-3d0288e2b5c2"). InnerVolumeSpecName "kube-api-access-8l6mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.602640 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f421ee4e-891c-4fff-b52a-3d0288e2b5c2" (UID: "f421ee4e-891c-4fff-b52a-3d0288e2b5c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.603093 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-config-data" (OuterVolumeSpecName: "config-data") pod "f421ee4e-891c-4fff-b52a-3d0288e2b5c2" (UID: "f421ee4e-891c-4fff-b52a-3d0288e2b5c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.631608 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f421ee4e-891c-4fff-b52a-3d0288e2b5c2" (UID: "f421ee4e-891c-4fff-b52a-3d0288e2b5c2"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.677242 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.677308 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.677321 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l6mx\" (UniqueName: \"kubernetes.io/projected/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-kube-api-access-8l6mx\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.677332 4715 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f421ee4e-891c-4fff-b52a-3d0288e2b5c2-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.816252 4715 generic.go:334] "Generic (PLEG): container finished" podID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" containerID="680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298" exitCode=0 Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.816305 4715 generic.go:334] "Generic (PLEG): container finished" podID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" containerID="f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104" exitCode=143 Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.816621 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.816614 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f421ee4e-891c-4fff-b52a-3d0288e2b5c2","Type":"ContainerDied","Data":"680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298"} Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.816788 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f421ee4e-891c-4fff-b52a-3d0288e2b5c2","Type":"ContainerDied","Data":"f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104"} Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.816822 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f421ee4e-891c-4fff-b52a-3d0288e2b5c2","Type":"ContainerDied","Data":"a5d28f65bf6e5172395be3e177247b07cb123ef73242893c561b43c5e49da529"} Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.816857 4715 scope.go:117] "RemoveContainer" containerID="680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.855510 4715 scope.go:117] "RemoveContainer" containerID="f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.881145 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.901390 4715 scope.go:117] "RemoveContainer" containerID="680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298" Dec 10 09:54:45 crc kubenswrapper[4715]: E1210 09:54:45.901765 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298\": container with ID starting with 680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298 not found: ID does not exist" containerID="680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.901804 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298"} err="failed to get container status \"680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298\": rpc error: code = NotFound desc = could not find container \"680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298\": container with ID starting with 680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298 not found: ID does not exist" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.901846 4715 scope.go:117] "RemoveContainer" containerID="f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104" Dec 10 09:54:45 crc kubenswrapper[4715]: E1210 09:54:45.902081 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104\": container with ID starting with f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104 not found: ID does not exist" containerID="f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.902131 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104"} err="failed to get container status \"f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104\": rpc error: code = NotFound desc = could not find container \"f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104\": container with ID starting with f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104 not found: ID does not exist" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.902144 4715 scope.go:117] "RemoveContainer" containerID="680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.902413 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298"} err="failed to get container status \"680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298\": rpc error: code = NotFound desc = could not find container \"680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298\": container with ID starting with 680ff009dae4be410d9e1e390a1f297761ed7de7bd39698e2b094ff1f7e5a298 not found: ID does not exist" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.902462 4715 scope.go:117] "RemoveContainer" containerID="f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.913196 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.918389 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104"} err="failed to get container status \"f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104\": rpc error: code = NotFound desc = could not find container \"f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104\": container with ID starting with f68eebcf7d13b2220b86421d4599ff02ceb5be3a56717afd7e128efc27169104 not found: ID does not exist" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.918461 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:45 crc kubenswrapper[4715]: E1210 09:54:45.919146 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" containerName="nova-metadata-log" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.919166 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" containerName="nova-metadata-log" Dec 10 09:54:45 crc kubenswrapper[4715]: E1210 09:54:45.919178 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" containerName="dnsmasq-dns" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.919203 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" containerName="dnsmasq-dns" Dec 10 09:54:45 crc kubenswrapper[4715]: E1210 09:54:45.919239 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9936c060-cb6f-4c92-895b-c0381d18085e" containerName="nova-manage" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.919246 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9936c060-cb6f-4c92-895b-c0381d18085e" containerName="nova-manage" Dec 10 09:54:45 crc kubenswrapper[4715]: E1210 09:54:45.919256 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" containerName="nova-metadata-metadata" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.919261 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" containerName="nova-metadata-metadata" Dec 10 09:54:45 crc kubenswrapper[4715]: E1210 09:54:45.919464 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" containerName="init" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.919472 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" containerName="init" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.921065 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0b9c91-b4e1-4112-ae4f-4a95bf0e2a25" containerName="dnsmasq-dns" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.921090 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" containerName="nova-metadata-log" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.921098 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9936c060-cb6f-4c92-895b-c0381d18085e" containerName="nova-manage" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.921114 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" containerName="nova-metadata-metadata" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.922294 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.927251 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.928775 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 09:54:45 crc kubenswrapper[4715]: I1210 09:54:45.946024 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.085489 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-logs\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.086250 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.086293 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-899vz\" (UniqueName: \"kubernetes.io/projected/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-kube-api-access-899vz\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.086348 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.086430 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-config-data\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.188265 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-logs\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.188311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.188327 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-899vz\" (UniqueName: \"kubernetes.io/projected/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-kube-api-access-899vz\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.188354 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.188398 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-config-data\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.189085 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-logs\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.201772 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.207513 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.208688 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-config-data\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.210760 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-899vz\" (UniqueName: \"kubernetes.io/projected/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-kube-api-access-899vz\") pod \"nova-metadata-0\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.282359 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.319385 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.493757 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-config-data\") pod \"af79009e-1855-4f22-8397-b1c762175b72\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.494171 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4rgm\" (UniqueName: \"kubernetes.io/projected/af79009e-1855-4f22-8397-b1c762175b72-kube-api-access-h4rgm\") pod \"af79009e-1855-4f22-8397-b1c762175b72\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.494260 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-combined-ca-bundle\") pod \"af79009e-1855-4f22-8397-b1c762175b72\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.494313 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-scripts\") pod \"af79009e-1855-4f22-8397-b1c762175b72\" (UID: \"af79009e-1855-4f22-8397-b1c762175b72\") " Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.499843 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af79009e-1855-4f22-8397-b1c762175b72-kube-api-access-h4rgm" (OuterVolumeSpecName: "kube-api-access-h4rgm") pod "af79009e-1855-4f22-8397-b1c762175b72" (UID: "af79009e-1855-4f22-8397-b1c762175b72"). InnerVolumeSpecName "kube-api-access-h4rgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.507326 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-scripts" (OuterVolumeSpecName: "scripts") pod "af79009e-1855-4f22-8397-b1c762175b72" (UID: "af79009e-1855-4f22-8397-b1c762175b72"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.530340 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af79009e-1855-4f22-8397-b1c762175b72" (UID: "af79009e-1855-4f22-8397-b1c762175b72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.531197 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.533387 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-config-data" (OuterVolumeSpecName: "config-data") pod "af79009e-1855-4f22-8397-b1c762175b72" (UID: "af79009e-1855-4f22-8397-b1c762175b72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.597117 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4rgm\" (UniqueName: \"kubernetes.io/projected/af79009e-1855-4f22-8397-b1c762175b72-kube-api-access-h4rgm\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.597156 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.597167 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.597179 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af79009e-1855-4f22-8397-b1c762175b72-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.698941 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-config-data\") pod \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.699409 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-combined-ca-bundle\") pod \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.699530 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsx5l\" (UniqueName: \"kubernetes.io/projected/cdba20b8-d67e-431f-bdcc-2763de1a1fea-kube-api-access-vsx5l\") pod \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\" (UID: \"cdba20b8-d67e-431f-bdcc-2763de1a1fea\") " Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.702312 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdba20b8-d67e-431f-bdcc-2763de1a1fea-kube-api-access-vsx5l" (OuterVolumeSpecName: "kube-api-access-vsx5l") pod "cdba20b8-d67e-431f-bdcc-2763de1a1fea" (UID: "cdba20b8-d67e-431f-bdcc-2763de1a1fea"). InnerVolumeSpecName "kube-api-access-vsx5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.723531 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-config-data" (OuterVolumeSpecName: "config-data") pod "cdba20b8-d67e-431f-bdcc-2763de1a1fea" (UID: "cdba20b8-d67e-431f-bdcc-2763de1a1fea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.723634 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdba20b8-d67e-431f-bdcc-2763de1a1fea" (UID: "cdba20b8-d67e-431f-bdcc-2763de1a1fea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.802056 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.802099 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdba20b8-d67e-431f-bdcc-2763de1a1fea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.802110 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsx5l\" (UniqueName: \"kubernetes.io/projected/cdba20b8-d67e-431f-bdcc-2763de1a1fea-kube-api-access-vsx5l\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:46 crc kubenswrapper[4715]: W1210 09:54:46.813286 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod378f8c2b_6bd7_4440_8a6c_49c76ebe9d0f.slice/crio-c788d1cd2c1f3e981aa5344f12667b5ec607e6cfd81fa9973c2556c5281242b7 WatchSource:0}: Error finding container c788d1cd2c1f3e981aa5344f12667b5ec607e6cfd81fa9973c2556c5281242b7: Status 404 returned error can't find the container with id c788d1cd2c1f3e981aa5344f12667b5ec607e6cfd81fa9973c2556c5281242b7 Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.815812 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.825665 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f","Type":"ContainerStarted","Data":"c788d1cd2c1f3e981aa5344f12667b5ec607e6cfd81fa9973c2556c5281242b7"} Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.827287 4715 generic.go:334] "Generic (PLEG): container finished" podID="cdba20b8-d67e-431f-bdcc-2763de1a1fea" containerID="a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23" exitCode=0 Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.827329 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.827361 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cdba20b8-d67e-431f-bdcc-2763de1a1fea","Type":"ContainerDied","Data":"a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23"} Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.827393 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cdba20b8-d67e-431f-bdcc-2763de1a1fea","Type":"ContainerDied","Data":"2fd32fdf2717004e6cc59df90c266d388cd34ec3be145964a20bf048df983f10"} Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.827413 4715 scope.go:117] "RemoveContainer" containerID="a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.860692 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-thxxv" event={"ID":"af79009e-1855-4f22-8397-b1c762175b72","Type":"ContainerDied","Data":"99fef4ee72eacf461fd9ea40182279fa22948f51ba3ca250bab9600da3cbdfc2"} Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.860767 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99fef4ee72eacf461fd9ea40182279fa22948f51ba3ca250bab9600da3cbdfc2" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.860863 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-thxxv" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.894214 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.895035 4715 scope.go:117] "RemoveContainer" containerID="a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23" Dec 10 09:54:46 crc kubenswrapper[4715]: E1210 09:54:46.898286 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23\": container with ID starting with a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23 not found: ID does not exist" containerID="a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.898367 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23"} err="failed to get container status \"a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23\": rpc error: code = NotFound desc = could not find container \"a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23\": container with ID starting with a75bb79647d7eba9ffcd2049943e4e083a2b8f421d4ba4381834d289f7d48b23 not found: ID does not exist" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.913341 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.923034 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:54:46 crc kubenswrapper[4715]: E1210 09:54:46.923692 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af79009e-1855-4f22-8397-b1c762175b72" containerName="nova-cell1-conductor-db-sync" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.923709 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="af79009e-1855-4f22-8397-b1c762175b72" containerName="nova-cell1-conductor-db-sync" Dec 10 09:54:46 crc kubenswrapper[4715]: E1210 09:54:46.923723 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdba20b8-d67e-431f-bdcc-2763de1a1fea" containerName="nova-scheduler-scheduler" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.923729 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdba20b8-d67e-431f-bdcc-2763de1a1fea" containerName="nova-scheduler-scheduler" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.923928 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdba20b8-d67e-431f-bdcc-2763de1a1fea" containerName="nova-scheduler-scheduler" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.923950 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="af79009e-1855-4f22-8397-b1c762175b72" containerName="nova-cell1-conductor-db-sync" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.924597 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.927895 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.933702 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.949503 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.950981 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.954321 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 10 09:54:46 crc kubenswrapper[4715]: I1210 09:54:46.958103 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.011789 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f9qt\" (UniqueName: \"kubernetes.io/projected/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-kube-api-access-9f9qt\") pod \"nova-scheduler-0\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.011877 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-config-data\") pod \"nova-scheduler-0\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.011931 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.114155 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7748551a-94b0-47c5-8360-584b386b0bf2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7748551a-94b0-47c5-8360-584b386b0bf2\") " pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.114237 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f9qt\" (UniqueName: \"kubernetes.io/projected/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-kube-api-access-9f9qt\") pod \"nova-scheduler-0\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.114309 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmg2p\" (UniqueName: \"kubernetes.io/projected/7748551a-94b0-47c5-8360-584b386b0bf2-kube-api-access-cmg2p\") pod \"nova-cell1-conductor-0\" (UID: \"7748551a-94b0-47c5-8360-584b386b0bf2\") " pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.114340 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-config-data\") pod \"nova-scheduler-0\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.114376 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.114443 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7748551a-94b0-47c5-8360-584b386b0bf2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7748551a-94b0-47c5-8360-584b386b0bf2\") " pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.121888 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.122457 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-config-data\") pod \"nova-scheduler-0\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.136572 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f9qt\" (UniqueName: \"kubernetes.io/projected/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-kube-api-access-9f9qt\") pod \"nova-scheduler-0\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " pod="openstack/nova-scheduler-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.216800 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7748551a-94b0-47c5-8360-584b386b0bf2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7748551a-94b0-47c5-8360-584b386b0bf2\") " pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.216978 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmg2p\" (UniqueName: \"kubernetes.io/projected/7748551a-94b0-47c5-8360-584b386b0bf2-kube-api-access-cmg2p\") pod \"nova-cell1-conductor-0\" (UID: \"7748551a-94b0-47c5-8360-584b386b0bf2\") " pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.217059 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7748551a-94b0-47c5-8360-584b386b0bf2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7748551a-94b0-47c5-8360-584b386b0bf2\") " pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.221494 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7748551a-94b0-47c5-8360-584b386b0bf2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7748551a-94b0-47c5-8360-584b386b0bf2\") " pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.221943 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7748551a-94b0-47c5-8360-584b386b0bf2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7748551a-94b0-47c5-8360-584b386b0bf2\") " pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.234188 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmg2p\" (UniqueName: \"kubernetes.io/projected/7748551a-94b0-47c5-8360-584b386b0bf2-kube-api-access-cmg2p\") pod \"nova-cell1-conductor-0\" (UID: \"7748551a-94b0-47c5-8360-584b386b0bf2\") " pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.312797 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.340462 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.511954 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.623237 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-combined-ca-bundle\") pod \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.623392 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgs2f\" (UniqueName: \"kubernetes.io/projected/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-kube-api-access-lgs2f\") pod \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.623492 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-logs\") pod \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.623561 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-config-data\") pod \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\" (UID: \"16c382ac-62d5-40b2-9fa3-fb6e57612f8a\") " Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.624926 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-logs" (OuterVolumeSpecName: "logs") pod "16c382ac-62d5-40b2-9fa3-fb6e57612f8a" (UID: "16c382ac-62d5-40b2-9fa3-fb6e57612f8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.630367 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-kube-api-access-lgs2f" (OuterVolumeSpecName: "kube-api-access-lgs2f") pod "16c382ac-62d5-40b2-9fa3-fb6e57612f8a" (UID: "16c382ac-62d5-40b2-9fa3-fb6e57612f8a"). InnerVolumeSpecName "kube-api-access-lgs2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.630769 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdba20b8-d67e-431f-bdcc-2763de1a1fea" path="/var/lib/kubelet/pods/cdba20b8-d67e-431f-bdcc-2763de1a1fea/volumes" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.631359 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f421ee4e-891c-4fff-b52a-3d0288e2b5c2" path="/var/lib/kubelet/pods/f421ee4e-891c-4fff-b52a-3d0288e2b5c2/volumes" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.659250 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-config-data" (OuterVolumeSpecName: "config-data") pod "16c382ac-62d5-40b2-9fa3-fb6e57612f8a" (UID: "16c382ac-62d5-40b2-9fa3-fb6e57612f8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.672938 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16c382ac-62d5-40b2-9fa3-fb6e57612f8a" (UID: "16c382ac-62d5-40b2-9fa3-fb6e57612f8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.728381 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgs2f\" (UniqueName: \"kubernetes.io/projected/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-kube-api-access-lgs2f\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.728420 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.728433 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.728444 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c382ac-62d5-40b2-9fa3-fb6e57612f8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:54:47 crc kubenswrapper[4715]: W1210 09:54:47.806105 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13438f95_4e2b_4225_afa0_0b2c13a9b4b2.slice/crio-1600f912e42970dab0543e5746c65164a07604fc45e05ba5cba1dd8b91015134 WatchSource:0}: Error finding container 1600f912e42970dab0543e5746c65164a07604fc45e05ba5cba1dd8b91015134: Status 404 returned error can't find the container with id 1600f912e42970dab0543e5746c65164a07604fc45e05ba5cba1dd8b91015134 Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.806439 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.871436 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"13438f95-4e2b-4225-afa0-0b2c13a9b4b2","Type":"ContainerStarted","Data":"1600f912e42970dab0543e5746c65164a07604fc45e05ba5cba1dd8b91015134"} Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.874305 4715 generic.go:334] "Generic (PLEG): container finished" podID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerID="9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c" exitCode=0 Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.874366 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.874413 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16c382ac-62d5-40b2-9fa3-fb6e57612f8a","Type":"ContainerDied","Data":"9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c"} Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.874470 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16c382ac-62d5-40b2-9fa3-fb6e57612f8a","Type":"ContainerDied","Data":"8ba9d0335c62e3af04fadff58e6d64fd1b55689b296ee81f9e11ba10d332986d"} Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.874500 4715 scope.go:117] "RemoveContainer" containerID="9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.882925 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f","Type":"ContainerStarted","Data":"9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d"} Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.882970 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f","Type":"ContainerStarted","Data":"d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5"} Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.917945 4715 scope.go:117] "RemoveContainer" containerID="d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a" Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.942263 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 09:54:47 crc kubenswrapper[4715]: I1210 09:54:47.954436 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.954418057 podStartE2EDuration="2.954418057s" podCreationTimestamp="2025-12-10 09:54:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:54:47.913425729 +0000 UTC m=+1250.656971980" watchObservedRunningTime="2025-12-10 09:54:47.954418057 +0000 UTC m=+1250.697964308" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.106787 4715 scope.go:117] "RemoveContainer" containerID="9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.107982 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:54:48 crc kubenswrapper[4715]: E1210 09:54:48.108347 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c\": container with ID starting with 9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c not found: ID does not exist" containerID="9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.108392 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c"} err="failed to get container status \"9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c\": rpc error: code = NotFound desc = could not find container \"9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c\": container with ID starting with 9a137b6af39a1ab90d3edda28a4dee5d34ff0c3c407cd8f98690d5e5b2f81a1c not found: ID does not exist" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.108420 4715 scope.go:117] "RemoveContainer" containerID="d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a" Dec 10 09:54:48 crc kubenswrapper[4715]: E1210 09:54:48.109412 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a\": container with ID starting with d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a not found: ID does not exist" containerID="d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.109541 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a"} err="failed to get container status \"d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a\": rpc error: code = NotFound desc = could not find container \"d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a\": container with ID starting with d67e10c91bbd190625096641cdd2c2933ce0fc784027eee01a4d1a64bbb3650a not found: ID does not exist" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.120622 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.177118 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 09:54:48 crc kubenswrapper[4715]: E1210 09:54:48.177792 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerName="nova-api-api" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.177812 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerName="nova-api-api" Dec 10 09:54:48 crc kubenswrapper[4715]: E1210 09:54:48.177841 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerName="nova-api-log" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.177849 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerName="nova-api-log" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.178190 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerName="nova-api-log" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.178214 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" containerName="nova-api-api" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.179270 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.186238 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.209013 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.348146 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.348236 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndl7t\" (UniqueName: \"kubernetes.io/projected/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-kube-api-access-ndl7t\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.348563 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-logs\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.348735 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-config-data\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.450671 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-logs\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.450736 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-config-data\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.450811 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.450865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndl7t\" (UniqueName: \"kubernetes.io/projected/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-kube-api-access-ndl7t\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.451193 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-logs\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.461636 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.461668 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-config-data\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.481011 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndl7t\" (UniqueName: \"kubernetes.io/projected/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-kube-api-access-ndl7t\") pod \"nova-api-0\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.504660 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.895569 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"13438f95-4e2b-4225-afa0-0b2c13a9b4b2","Type":"ContainerStarted","Data":"53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80"} Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.924827 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7748551a-94b0-47c5-8360-584b386b0bf2","Type":"ContainerStarted","Data":"92418b91af909feea91b639843e1039d80c2a2890a59d9e7519aca6b7aaf63c1"} Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.924884 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7748551a-94b0-47c5-8360-584b386b0bf2","Type":"ContainerStarted","Data":"44729aee7abb06346de8897ff7ca404b03a29829a30500b85677c641713e0b0f"} Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.956898 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.956874782 podStartE2EDuration="2.956874782s" podCreationTimestamp="2025-12-10 09:54:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:54:48.921066321 +0000 UTC m=+1251.664612592" watchObservedRunningTime="2025-12-10 09:54:48.956874782 +0000 UTC m=+1251.700421033" Dec 10 09:54:48 crc kubenswrapper[4715]: I1210 09:54:48.959169 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.959147767 podStartE2EDuration="2.959147767s" podCreationTimestamp="2025-12-10 09:54:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:54:48.947071083 +0000 UTC m=+1251.690617394" watchObservedRunningTime="2025-12-10 09:54:48.959147767 +0000 UTC m=+1251.702694018" Dec 10 09:54:49 crc kubenswrapper[4715]: I1210 09:54:49.022553 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:54:49 crc kubenswrapper[4715]: I1210 09:54:49.621436 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16c382ac-62d5-40b2-9fa3-fb6e57612f8a" path="/var/lib/kubelet/pods/16c382ac-62d5-40b2-9fa3-fb6e57612f8a/volumes" Dec 10 09:54:49 crc kubenswrapper[4715]: I1210 09:54:49.936431 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b67d98ea-edd4-426b-96f9-b8919c5c1ceb","Type":"ContainerStarted","Data":"7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c"} Dec 10 09:54:49 crc kubenswrapper[4715]: I1210 09:54:49.936476 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b67d98ea-edd4-426b-96f9-b8919c5c1ceb","Type":"ContainerStarted","Data":"ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3"} Dec 10 09:54:49 crc kubenswrapper[4715]: I1210 09:54:49.936491 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b67d98ea-edd4-426b-96f9-b8919c5c1ceb","Type":"ContainerStarted","Data":"50d0b7b32b697965b2d8238c2f4beb28bcaa30cb3ce53a61ad07af20da9711d3"} Dec 10 09:54:49 crc kubenswrapper[4715]: I1210 09:54:49.937001 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:49 crc kubenswrapper[4715]: I1210 09:54:49.963398 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.9633786519999998 podStartE2EDuration="1.963378652s" podCreationTimestamp="2025-12-10 09:54:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:54:49.959203563 +0000 UTC m=+1252.702749814" watchObservedRunningTime="2025-12-10 09:54:49.963378652 +0000 UTC m=+1252.706924903" Dec 10 09:54:51 crc kubenswrapper[4715]: I1210 09:54:51.282682 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 09:54:51 crc kubenswrapper[4715]: I1210 09:54:51.282751 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 09:54:52 crc kubenswrapper[4715]: I1210 09:54:52.313484 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 09:54:56 crc kubenswrapper[4715]: I1210 09:54:56.283550 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 09:54:56 crc kubenswrapper[4715]: I1210 09:54:56.283997 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 09:54:57 crc kubenswrapper[4715]: I1210 09:54:57.300157 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 09:54:57 crc kubenswrapper[4715]: I1210 09:54:57.300157 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 09:54:57 crc kubenswrapper[4715]: I1210 09:54:57.313376 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 09:54:57 crc kubenswrapper[4715]: I1210 09:54:57.342625 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 09:54:57 crc kubenswrapper[4715]: I1210 09:54:57.380227 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 10 09:54:58 crc kubenswrapper[4715]: I1210 09:54:58.041137 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 09:54:58 crc kubenswrapper[4715]: I1210 09:54:58.505297 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 09:54:58 crc kubenswrapper[4715]: I1210 09:54:58.505621 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 09:54:59 crc kubenswrapper[4715]: I1210 09:54:59.589125 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 09:54:59 crc kubenswrapper[4715]: I1210 09:54:59.589125 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 09:54:59 crc kubenswrapper[4715]: I1210 09:54:59.832894 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 09:55:06 crc kubenswrapper[4715]: I1210 09:55:06.289578 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 09:55:06 crc kubenswrapper[4715]: I1210 09:55:06.290449 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 09:55:06 crc kubenswrapper[4715]: I1210 09:55:06.294412 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 09:55:06 crc kubenswrapper[4715]: I1210 09:55:06.295633 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 09:55:06 crc kubenswrapper[4715]: I1210 09:55:06.979012 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.129641 4715 generic.go:334] "Generic (PLEG): container finished" podID="aaad46b8-fed6-4ed4-9845-5ddfa954edad" containerID="34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2" exitCode=137 Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.129717 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.129741 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"aaad46b8-fed6-4ed4-9845-5ddfa954edad","Type":"ContainerDied","Data":"34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2"} Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.129782 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"aaad46b8-fed6-4ed4-9845-5ddfa954edad","Type":"ContainerDied","Data":"409ea07dc02cfbffddc4bafb9b667fb5d3a7858b6adac1bb0b290a3f0b6d172a"} Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.129801 4715 scope.go:117] "RemoveContainer" containerID="34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.149557 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrblp\" (UniqueName: \"kubernetes.io/projected/aaad46b8-fed6-4ed4-9845-5ddfa954edad-kube-api-access-jrblp\") pod \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.149700 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-config-data\") pod \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.149720 4715 scope.go:117] "RemoveContainer" containerID="34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.149801 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-combined-ca-bundle\") pod \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\" (UID: \"aaad46b8-fed6-4ed4-9845-5ddfa954edad\") " Dec 10 09:55:07 crc kubenswrapper[4715]: E1210 09:55:07.150559 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2\": container with ID starting with 34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2 not found: ID does not exist" containerID="34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.150607 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2"} err="failed to get container status \"34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2\": rpc error: code = NotFound desc = could not find container \"34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2\": container with ID starting with 34f2ba5f4e7ad2d29cd8d40301e3f321719f7f5da443b3468c6da5d815eda8b2 not found: ID does not exist" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.156159 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaad46b8-fed6-4ed4-9845-5ddfa954edad-kube-api-access-jrblp" (OuterVolumeSpecName: "kube-api-access-jrblp") pod "aaad46b8-fed6-4ed4-9845-5ddfa954edad" (UID: "aaad46b8-fed6-4ed4-9845-5ddfa954edad"). InnerVolumeSpecName "kube-api-access-jrblp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.176583 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-config-data" (OuterVolumeSpecName: "config-data") pod "aaad46b8-fed6-4ed4-9845-5ddfa954edad" (UID: "aaad46b8-fed6-4ed4-9845-5ddfa954edad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.180705 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aaad46b8-fed6-4ed4-9845-5ddfa954edad" (UID: "aaad46b8-fed6-4ed4-9845-5ddfa954edad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.252187 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrblp\" (UniqueName: \"kubernetes.io/projected/aaad46b8-fed6-4ed4-9845-5ddfa954edad-kube-api-access-jrblp\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.252232 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.252243 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaad46b8-fed6-4ed4-9845-5ddfa954edad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.469017 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.483137 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.496987 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 09:55:07 crc kubenswrapper[4715]: E1210 09:55:07.497499 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaad46b8-fed6-4ed4-9845-5ddfa954edad" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.497519 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaad46b8-fed6-4ed4-9845-5ddfa954edad" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.497713 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaad46b8-fed6-4ed4-9845-5ddfa954edad" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.498408 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.501264 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.501562 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.501989 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.512062 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.623229 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaad46b8-fed6-4ed4-9845-5ddfa954edad" path="/var/lib/kubelet/pods/aaad46b8-fed6-4ed4-9845-5ddfa954edad/volumes" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.658937 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.659000 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.659046 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.659110 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.659552 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99jpx\" (UniqueName: \"kubernetes.io/projected/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-kube-api-access-99jpx\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.761459 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99jpx\" (UniqueName: \"kubernetes.io/projected/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-kube-api-access-99jpx\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.761539 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.761569 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.761593 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.761610 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.767105 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.767198 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.776362 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.776891 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.778719 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99jpx\" (UniqueName: \"kubernetes.io/projected/87397cd7-eabb-4b0b-809a-c9f2ab53e8be-kube-api-access-99jpx\") pod \"nova-cell1-novncproxy-0\" (UID: \"87397cd7-eabb-4b0b-809a-c9f2ab53e8be\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:07 crc kubenswrapper[4715]: I1210 09:55:07.832386 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:08 crc kubenswrapper[4715]: I1210 09:55:08.284040 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 09:55:08 crc kubenswrapper[4715]: W1210 09:55:08.289759 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87397cd7_eabb_4b0b_809a_c9f2ab53e8be.slice/crio-80db51648032c571b792abbee711c9ab222a4ea36b6b67d831ab8f0ca1c48258 WatchSource:0}: Error finding container 80db51648032c571b792abbee711c9ab222a4ea36b6b67d831ab8f0ca1c48258: Status 404 returned error can't find the container with id 80db51648032c571b792abbee711c9ab222a4ea36b6b67d831ab8f0ca1c48258 Dec 10 09:55:08 crc kubenswrapper[4715]: I1210 09:55:08.509650 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 09:55:08 crc kubenswrapper[4715]: I1210 09:55:08.510434 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 09:55:08 crc kubenswrapper[4715]: I1210 09:55:08.516908 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 09:55:08 crc kubenswrapper[4715]: I1210 09:55:08.521363 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.169045 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"87397cd7-eabb-4b0b-809a-c9f2ab53e8be","Type":"ContainerStarted","Data":"1b67d4848f01105da08ff5101862175d396ccce20864f46fa8f9b9d9d9d78247"} Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.169657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"87397cd7-eabb-4b0b-809a-c9f2ab53e8be","Type":"ContainerStarted","Data":"80db51648032c571b792abbee711c9ab222a4ea36b6b67d831ab8f0ca1c48258"} Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.170453 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.174774 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.226817 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.226788336 podStartE2EDuration="2.226788336s" podCreationTimestamp="2025-12-10 09:55:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:55:09.203261905 +0000 UTC m=+1271.946808156" watchObservedRunningTime="2025-12-10 09:55:09.226788336 +0000 UTC m=+1271.970334587" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.422644 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fcqpx"] Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.428380 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.442219 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fcqpx"] Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.509305 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.509436 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.509479 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwhvd\" (UniqueName: \"kubernetes.io/projected/a7490f58-8a80-453b-924b-3f14aeceeff8-kube-api-access-fwhvd\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.509558 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.509626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-config\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.509681 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.611647 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwhvd\" (UniqueName: \"kubernetes.io/projected/a7490f58-8a80-453b-924b-3f14aeceeff8-kube-api-access-fwhvd\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.611711 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.611740 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-config\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.611774 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.611835 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.611883 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.613786 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-config\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.613836 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.617534 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.618091 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.618316 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.642963 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwhvd\" (UniqueName: \"kubernetes.io/projected/a7490f58-8a80-453b-924b-3f14aeceeff8-kube-api-access-fwhvd\") pod \"dnsmasq-dns-89c5cd4d5-fcqpx\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:09 crc kubenswrapper[4715]: I1210 09:55:09.839361 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:10 crc kubenswrapper[4715]: I1210 09:55:10.361324 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fcqpx"] Dec 10 09:55:10 crc kubenswrapper[4715]: W1210 09:55:10.361720 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7490f58_8a80_453b_924b_3f14aeceeff8.slice/crio-40c786517981b7605b57118882b6fef08690177aca913f5108808fb0871d09bc WatchSource:0}: Error finding container 40c786517981b7605b57118882b6fef08690177aca913f5108808fb0871d09bc: Status 404 returned error can't find the container with id 40c786517981b7605b57118882b6fef08690177aca913f5108808fb0871d09bc Dec 10 09:55:11 crc kubenswrapper[4715]: I1210 09:55:11.199032 4715 generic.go:334] "Generic (PLEG): container finished" podID="a7490f58-8a80-453b-924b-3f14aeceeff8" containerID="5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837" exitCode=0 Dec 10 09:55:11 crc kubenswrapper[4715]: I1210 09:55:11.199113 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" event={"ID":"a7490f58-8a80-453b-924b-3f14aeceeff8","Type":"ContainerDied","Data":"5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837"} Dec 10 09:55:11 crc kubenswrapper[4715]: I1210 09:55:11.199419 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" event={"ID":"a7490f58-8a80-453b-924b-3f14aeceeff8","Type":"ContainerStarted","Data":"40c786517981b7605b57118882b6fef08690177aca913f5108808fb0871d09bc"} Dec 10 09:55:11 crc kubenswrapper[4715]: I1210 09:55:11.408345 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:55:11 crc kubenswrapper[4715]: I1210 09:55:11.408948 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="ceilometer-central-agent" containerID="cri-o://c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c" gracePeriod=30 Dec 10 09:55:11 crc kubenswrapper[4715]: I1210 09:55:11.408973 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="proxy-httpd" containerID="cri-o://a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3" gracePeriod=30 Dec 10 09:55:11 crc kubenswrapper[4715]: I1210 09:55:11.409019 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="sg-core" containerID="cri-o://1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491" gracePeriod=30 Dec 10 09:55:11 crc kubenswrapper[4715]: I1210 09:55:11.409082 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="ceilometer-notification-agent" containerID="cri-o://98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b" gracePeriod=30 Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.077983 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.211992 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" event={"ID":"a7490f58-8a80-453b-924b-3f14aeceeff8","Type":"ContainerStarted","Data":"031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927"} Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.213248 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.217060 4715 generic.go:334] "Generic (PLEG): container finished" podID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerID="a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3" exitCode=0 Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.217090 4715 generic.go:334] "Generic (PLEG): container finished" podID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerID="1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491" exitCode=2 Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.217100 4715 generic.go:334] "Generic (PLEG): container finished" podID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerID="c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c" exitCode=0 Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.217095 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f30313c-cf6f-4ec5-a2da-70c9f7575918","Type":"ContainerDied","Data":"a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3"} Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.217144 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f30313c-cf6f-4ec5-a2da-70c9f7575918","Type":"ContainerDied","Data":"1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491"} Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.217161 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f30313c-cf6f-4ec5-a2da-70c9f7575918","Type":"ContainerDied","Data":"c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c"} Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.217291 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerName="nova-api-log" containerID="cri-o://ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3" gracePeriod=30 Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.217340 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerName="nova-api-api" containerID="cri-o://7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c" gracePeriod=30 Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.241860 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" podStartSLOduration=3.241836738 podStartE2EDuration="3.241836738s" podCreationTimestamp="2025-12-10 09:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:55:12.234733036 +0000 UTC m=+1274.978279297" watchObservedRunningTime="2025-12-10 09:55:12.241836738 +0000 UTC m=+1274.985382989" Dec 10 09:55:12 crc kubenswrapper[4715]: I1210 09:55:12.833407 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:13 crc kubenswrapper[4715]: I1210 09:55:13.236543 4715 generic.go:334] "Generic (PLEG): container finished" podID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerID="ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3" exitCode=143 Dec 10 09:55:13 crc kubenswrapper[4715]: I1210 09:55:13.236833 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b67d98ea-edd4-426b-96f9-b8919c5c1ceb","Type":"ContainerDied","Data":"ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3"} Dec 10 09:55:15 crc kubenswrapper[4715]: I1210 09:55:15.837055 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:55:15 crc kubenswrapper[4715]: I1210 09:55:15.954060 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-combined-ca-bundle\") pod \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " Dec 10 09:55:15 crc kubenswrapper[4715]: I1210 09:55:15.954194 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndl7t\" (UniqueName: \"kubernetes.io/projected/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-kube-api-access-ndl7t\") pod \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " Dec 10 09:55:15 crc kubenswrapper[4715]: I1210 09:55:15.954229 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-config-data\") pod \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " Dec 10 09:55:15 crc kubenswrapper[4715]: I1210 09:55:15.954326 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-logs\") pod \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\" (UID: \"b67d98ea-edd4-426b-96f9-b8919c5c1ceb\") " Dec 10 09:55:15 crc kubenswrapper[4715]: I1210 09:55:15.955234 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-logs" (OuterVolumeSpecName: "logs") pod "b67d98ea-edd4-426b-96f9-b8919c5c1ceb" (UID: "b67d98ea-edd4-426b-96f9-b8919c5c1ceb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:55:15 crc kubenswrapper[4715]: I1210 09:55:15.991165 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-kube-api-access-ndl7t" (OuterVolumeSpecName: "kube-api-access-ndl7t") pod "b67d98ea-edd4-426b-96f9-b8919c5c1ceb" (UID: "b67d98ea-edd4-426b-96f9-b8919c5c1ceb"). InnerVolumeSpecName "kube-api-access-ndl7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:55:15 crc kubenswrapper[4715]: I1210 09:55:15.995101 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-config-data" (OuterVolumeSpecName: "config-data") pod "b67d98ea-edd4-426b-96f9-b8919c5c1ceb" (UID: "b67d98ea-edd4-426b-96f9-b8919c5c1ceb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.030495 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b67d98ea-edd4-426b-96f9-b8919c5c1ceb" (UID: "b67d98ea-edd4-426b-96f9-b8919c5c1ceb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.058163 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.058200 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.058212 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndl7t\" (UniqueName: \"kubernetes.io/projected/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-kube-api-access-ndl7t\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.058220 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67d98ea-edd4-426b-96f9-b8919c5c1ceb-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.090569 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.260316 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-log-httpd\") pod \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.260472 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-run-httpd\") pod \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.260571 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-scripts\") pod \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.260642 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-config-data\") pod \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.260744 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-combined-ca-bundle\") pod \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.260776 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-sg-core-conf-yaml\") pod \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.260850 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzbvj\" (UniqueName: \"kubernetes.io/projected/3f30313c-cf6f-4ec5-a2da-70c9f7575918-kube-api-access-jzbvj\") pod \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.260847 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3f30313c-cf6f-4ec5-a2da-70c9f7575918" (UID: "3f30313c-cf6f-4ec5-a2da-70c9f7575918"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.260880 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3f30313c-cf6f-4ec5-a2da-70c9f7575918" (UID: "3f30313c-cf6f-4ec5-a2da-70c9f7575918"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.260900 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-ceilometer-tls-certs\") pod \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\" (UID: \"3f30313c-cf6f-4ec5-a2da-70c9f7575918\") " Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.261970 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.261999 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f30313c-cf6f-4ec5-a2da-70c9f7575918-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.264545 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-scripts" (OuterVolumeSpecName: "scripts") pod "3f30313c-cf6f-4ec5-a2da-70c9f7575918" (UID: "3f30313c-cf6f-4ec5-a2da-70c9f7575918"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.264659 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f30313c-cf6f-4ec5-a2da-70c9f7575918-kube-api-access-jzbvj" (OuterVolumeSpecName: "kube-api-access-jzbvj") pod "3f30313c-cf6f-4ec5-a2da-70c9f7575918" (UID: "3f30313c-cf6f-4ec5-a2da-70c9f7575918"). InnerVolumeSpecName "kube-api-access-jzbvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.273561 4715 generic.go:334] "Generic (PLEG): container finished" podID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerID="98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b" exitCode=0 Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.273631 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f30313c-cf6f-4ec5-a2da-70c9f7575918","Type":"ContainerDied","Data":"98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b"} Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.273663 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f30313c-cf6f-4ec5-a2da-70c9f7575918","Type":"ContainerDied","Data":"973035c580cd48cc8c345c4d7875e312558bcea9aacf89efe07cd671cf39ba92"} Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.273684 4715 scope.go:117] "RemoveContainer" containerID="a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.273849 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.284106 4715 generic.go:334] "Generic (PLEG): container finished" podID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerID="7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c" exitCode=0 Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.284162 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b67d98ea-edd4-426b-96f9-b8919c5c1ceb","Type":"ContainerDied","Data":"7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c"} Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.284195 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b67d98ea-edd4-426b-96f9-b8919c5c1ceb","Type":"ContainerDied","Data":"50d0b7b32b697965b2d8238c2f4beb28bcaa30cb3ce53a61ad07af20da9711d3"} Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.284166 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.293361 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3f30313c-cf6f-4ec5-a2da-70c9f7575918" (UID: "3f30313c-cf6f-4ec5-a2da-70c9f7575918"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.309611 4715 scope.go:117] "RemoveContainer" containerID="1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.326610 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3f30313c-cf6f-4ec5-a2da-70c9f7575918" (UID: "3f30313c-cf6f-4ec5-a2da-70c9f7575918"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.337702 4715 scope.go:117] "RemoveContainer" containerID="98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.355115 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.363965 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.364003 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.364015 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzbvj\" (UniqueName: \"kubernetes.io/projected/3f30313c-cf6f-4ec5-a2da-70c9f7575918-kube-api-access-jzbvj\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.364028 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.365118 4715 scope.go:117] "RemoveContainer" containerID="c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.375309 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.375336 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f30313c-cf6f-4ec5-a2da-70c9f7575918" (UID: "3f30313c-cf6f-4ec5-a2da-70c9f7575918"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.389815 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.390572 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="ceilometer-central-agent" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.390653 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="ceilometer-central-agent" Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.390731 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="sg-core" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.390800 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="sg-core" Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.390871 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="proxy-httpd" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.390962 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="proxy-httpd" Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.391041 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerName="nova-api-api" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.391100 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerName="nova-api-api" Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.391160 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="ceilometer-notification-agent" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.391213 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="ceilometer-notification-agent" Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.391271 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerName="nova-api-log" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.391333 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerName="nova-api-log" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.391616 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerName="nova-api-log" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.391690 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" containerName="nova-api-api" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.391749 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="sg-core" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.391829 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="ceilometer-central-agent" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.391904 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="ceilometer-notification-agent" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.392091 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" containerName="proxy-httpd" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.392564 4715 scope.go:117] "RemoveContainer" containerID="a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.395657 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.397209 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3\": container with ID starting with a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3 not found: ID does not exist" containerID="a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.397249 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3"} err="failed to get container status \"a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3\": rpc error: code = NotFound desc = could not find container \"a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3\": container with ID starting with a374969c1a4fa110714507c6308e1c6fff3c5cb9b0b27b5d340e03dab65a21c3 not found: ID does not exist" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.397277 4715 scope.go:117] "RemoveContainer" containerID="1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491" Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.398349 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491\": container with ID starting with 1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491 not found: ID does not exist" containerID="1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.398458 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491"} err="failed to get container status \"1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491\": rpc error: code = NotFound desc = could not find container \"1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491\": container with ID starting with 1d2f541cb29fdcdd2b07b5e7cb17640751c93ef13e576d66250c205a24cad491 not found: ID does not exist" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.398572 4715 scope.go:117] "RemoveContainer" containerID="98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b" Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.399168 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b\": container with ID starting with 98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b not found: ID does not exist" containerID="98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.399202 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b"} err="failed to get container status \"98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b\": rpc error: code = NotFound desc = could not find container \"98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b\": container with ID starting with 98baa29018b8444bc53a77ede9bf85adf2c53c220c0c3ae0b04575cf1f9caa2b not found: ID does not exist" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.399218 4715 scope.go:117] "RemoveContainer" containerID="c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.400112 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.400369 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.401201 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.402145 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.403350 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c\": container with ID starting with c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c not found: ID does not exist" containerID="c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.403409 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c"} err="failed to get container status \"c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c\": rpc error: code = NotFound desc = could not find container \"c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c\": container with ID starting with c45513032d7a8ec6e6e39d10961c86323338c4ac6e2bf6dc93f81139f15f388c not found: ID does not exist" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.403441 4715 scope.go:117] "RemoveContainer" containerID="7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.415559 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-config-data" (OuterVolumeSpecName: "config-data") pod "3f30313c-cf6f-4ec5-a2da-70c9f7575918" (UID: "3f30313c-cf6f-4ec5-a2da-70c9f7575918"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.428213 4715 scope.go:117] "RemoveContainer" containerID="ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.449582 4715 scope.go:117] "RemoveContainer" containerID="7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c" Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.450162 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c\": container with ID starting with 7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c not found: ID does not exist" containerID="7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.450203 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c"} err="failed to get container status \"7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c\": rpc error: code = NotFound desc = could not find container \"7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c\": container with ID starting with 7b756f3357930a9d3929f43cbd681c2e972479a12d10d25551802a8da4006c9c not found: ID does not exist" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.450233 4715 scope.go:117] "RemoveContainer" containerID="ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3" Dec 10 09:55:16 crc kubenswrapper[4715]: E1210 09:55:16.450677 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3\": container with ID starting with ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3 not found: ID does not exist" containerID="ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.450719 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3"} err="failed to get container status \"ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3\": rpc error: code = NotFound desc = could not find container \"ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3\": container with ID starting with ed87ce4b4d90579a6fc391bdf34f51a16f01db3df1f6567682fe5497906f59a3 not found: ID does not exist" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.466264 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.466296 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f30313c-cf6f-4ec5-a2da-70c9f7575918-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.568355 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-config-data\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.568417 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7sqx\" (UniqueName: \"kubernetes.io/projected/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-kube-api-access-q7sqx\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.568440 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.568500 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-logs\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.568524 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-public-tls-certs\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.568579 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.603903 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.615156 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.642542 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.645412 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.648224 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.648224 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.648995 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.661259 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.670688 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-config-data\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.670759 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7sqx\" (UniqueName: \"kubernetes.io/projected/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-kube-api-access-q7sqx\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.670788 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.670859 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-logs\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.670888 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-public-tls-certs\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.670936 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.671519 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-logs\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.677348 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-config-data\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.677901 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.678613 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-public-tls-certs\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.690806 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.693660 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7sqx\" (UniqueName: \"kubernetes.io/projected/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-kube-api-access-q7sqx\") pod \"nova-api-0\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.723837 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.772793 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.773572 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfd7s\" (UniqueName: \"kubernetes.io/projected/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-kube-api-access-dfd7s\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.773643 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-config-data\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.773791 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.773896 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.774172 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-log-httpd\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.774307 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-scripts\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.774325 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-run-httpd\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.876216 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.876287 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfd7s\" (UniqueName: \"kubernetes.io/projected/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-kube-api-access-dfd7s\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.876326 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-config-data\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.876400 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.876415 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.876440 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-log-httpd\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.876472 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-scripts\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.876487 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-run-httpd\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.877149 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-run-httpd\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.877573 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-log-httpd\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.882393 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.882951 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-scripts\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.884545 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.887167 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-config-data\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.894868 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.895112 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfd7s\" (UniqueName: \"kubernetes.io/projected/3a26e430-a5f3-40cf-97fc-be1e01adb4e9-kube-api-access-dfd7s\") pod \"ceilometer-0\" (UID: \"3a26e430-a5f3-40cf-97fc-be1e01adb4e9\") " pod="openstack/ceilometer-0" Dec 10 09:55:16 crc kubenswrapper[4715]: I1210 09:55:16.965092 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 09:55:17 crc kubenswrapper[4715]: I1210 09:55:17.195267 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:17 crc kubenswrapper[4715]: I1210 09:55:17.294039 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c","Type":"ContainerStarted","Data":"69b2c45d467bd786fa882f5cf842583639f3ec55fdf69e33879f5b4768f81cdd"} Dec 10 09:55:17 crc kubenswrapper[4715]: I1210 09:55:17.461546 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 09:55:17 crc kubenswrapper[4715]: I1210 09:55:17.620996 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f30313c-cf6f-4ec5-a2da-70c9f7575918" path="/var/lib/kubelet/pods/3f30313c-cf6f-4ec5-a2da-70c9f7575918/volumes" Dec 10 09:55:17 crc kubenswrapper[4715]: I1210 09:55:17.622441 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b67d98ea-edd4-426b-96f9-b8919c5c1ceb" path="/var/lib/kubelet/pods/b67d98ea-edd4-426b-96f9-b8919c5c1ceb/volumes" Dec 10 09:55:17 crc kubenswrapper[4715]: I1210 09:55:17.834155 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:17 crc kubenswrapper[4715]: I1210 09:55:17.856777 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.307992 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c","Type":"ContainerStarted","Data":"e58beba0294f8367482faeb2cd43e629cd66839116f13421d18b41aa9e4b5a2d"} Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.308261 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c","Type":"ContainerStarted","Data":"504cbd12ed32a575d14a2d5f2c18bd1986a8723d9f51b3718de51edef6c82de9"} Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.315092 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a26e430-a5f3-40cf-97fc-be1e01adb4e9","Type":"ContainerStarted","Data":"2728e2129918aa49962373fe45c26b8fa3d43f24303786c23244645a58268cd5"} Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.315135 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a26e430-a5f3-40cf-97fc-be1e01adb4e9","Type":"ContainerStarted","Data":"9a43b8e8cecdda2bcd39de3031b226d34efe474adafb03d1f6fad0c9d6ef7154"} Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.332720 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.3327001259999998 podStartE2EDuration="2.332700126s" podCreationTimestamp="2025-12-10 09:55:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:55:18.32758955 +0000 UTC m=+1281.071135801" watchObservedRunningTime="2025-12-10 09:55:18.332700126 +0000 UTC m=+1281.076246377" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.336037 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.498626 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-xgcdc"] Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.499986 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.502460 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.502499 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.519134 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-xgcdc"] Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.615258 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.615451 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-scripts\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.615514 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6mpq\" (UniqueName: \"kubernetes.io/projected/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-kube-api-access-p6mpq\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.615592 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-config-data\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.717064 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-scripts\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.717117 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6mpq\" (UniqueName: \"kubernetes.io/projected/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-kube-api-access-p6mpq\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.717150 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-config-data\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.717234 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.722009 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-config-data\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.722295 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.723759 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-scripts\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.740659 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6mpq\" (UniqueName: \"kubernetes.io/projected/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-kube-api-access-p6mpq\") pod \"nova-cell1-cell-mapping-xgcdc\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:18 crc kubenswrapper[4715]: I1210 09:55:18.818795 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:19 crc kubenswrapper[4715]: I1210 09:55:19.315583 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-xgcdc"] Dec 10 09:55:19 crc kubenswrapper[4715]: I1210 09:55:19.335660 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a26e430-a5f3-40cf-97fc-be1e01adb4e9","Type":"ContainerStarted","Data":"97a9c56161bca4f82bc383bea8b7b9a7b2cd0116bbde05ffdb00617b698ea85b"} Dec 10 09:55:19 crc kubenswrapper[4715]: I1210 09:55:19.841573 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:55:19 crc kubenswrapper[4715]: I1210 09:55:19.927285 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q6kwd"] Dec 10 09:55:19 crc kubenswrapper[4715]: I1210 09:55:19.927599 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" podUID="7742e050-7183-4c20-9367-618075abadc8" containerName="dnsmasq-dns" containerID="cri-o://776f87f56130cb887230b8f13542b66aed7e25ffb3cb49306209bd93f96e9ca7" gracePeriod=10 Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.346695 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a26e430-a5f3-40cf-97fc-be1e01adb4e9","Type":"ContainerStarted","Data":"40cc5700d1b9c8dbb676467ad4f51411838434af821fa1374ba19d971458d815"} Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.352666 4715 generic.go:334] "Generic (PLEG): container finished" podID="7742e050-7183-4c20-9367-618075abadc8" containerID="776f87f56130cb887230b8f13542b66aed7e25ffb3cb49306209bd93f96e9ca7" exitCode=0 Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.352738 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" event={"ID":"7742e050-7183-4c20-9367-618075abadc8","Type":"ContainerDied","Data":"776f87f56130cb887230b8f13542b66aed7e25ffb3cb49306209bd93f96e9ca7"} Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.354551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xgcdc" event={"ID":"f07f968f-58f0-46c6-ac4e-d0cdcb966d10","Type":"ContainerStarted","Data":"69f349fcb33c1d2a5a8883c637c243d58b1b07882c0e4513dd824035be60d472"} Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.354595 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xgcdc" event={"ID":"f07f968f-58f0-46c6-ac4e-d0cdcb966d10","Type":"ContainerStarted","Data":"63e2bd24dfc5b44396dd9a08390b0b35fe49719a734ba2b5338f164d370f01d1"} Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.398616 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-xgcdc" podStartSLOduration=2.398599514 podStartE2EDuration="2.398599514s" podCreationTimestamp="2025-12-10 09:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:55:20.391852701 +0000 UTC m=+1283.135398962" watchObservedRunningTime="2025-12-10 09:55:20.398599514 +0000 UTC m=+1283.142145765" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.467474 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.653276 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-swift-storage-0\") pod \"7742e050-7183-4c20-9367-618075abadc8\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.653382 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46jkl\" (UniqueName: \"kubernetes.io/projected/7742e050-7183-4c20-9367-618075abadc8-kube-api-access-46jkl\") pod \"7742e050-7183-4c20-9367-618075abadc8\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.653507 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-sb\") pod \"7742e050-7183-4c20-9367-618075abadc8\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.653602 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-nb\") pod \"7742e050-7183-4c20-9367-618075abadc8\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.653650 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-svc\") pod \"7742e050-7183-4c20-9367-618075abadc8\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.653705 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-config\") pod \"7742e050-7183-4c20-9367-618075abadc8\" (UID: \"7742e050-7183-4c20-9367-618075abadc8\") " Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.661521 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7742e050-7183-4c20-9367-618075abadc8-kube-api-access-46jkl" (OuterVolumeSpecName: "kube-api-access-46jkl") pod "7742e050-7183-4c20-9367-618075abadc8" (UID: "7742e050-7183-4c20-9367-618075abadc8"). InnerVolumeSpecName "kube-api-access-46jkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.745681 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7742e050-7183-4c20-9367-618075abadc8" (UID: "7742e050-7183-4c20-9367-618075abadc8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.751623 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7742e050-7183-4c20-9367-618075abadc8" (UID: "7742e050-7183-4c20-9367-618075abadc8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.755890 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7742e050-7183-4c20-9367-618075abadc8" (UID: "7742e050-7183-4c20-9367-618075abadc8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.756138 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.756166 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46jkl\" (UniqueName: \"kubernetes.io/projected/7742e050-7183-4c20-9367-618075abadc8-kube-api-access-46jkl\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.756180 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.756192 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.768510 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7742e050-7183-4c20-9367-618075abadc8" (UID: "7742e050-7183-4c20-9367-618075abadc8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.811721 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-config" (OuterVolumeSpecName: "config") pod "7742e050-7183-4c20-9367-618075abadc8" (UID: "7742e050-7183-4c20-9367-618075abadc8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.858372 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:20 crc kubenswrapper[4715]: I1210 09:55:20.858412 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7742e050-7183-4c20-9367-618075abadc8-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:21 crc kubenswrapper[4715]: I1210 09:55:21.367425 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a26e430-a5f3-40cf-97fc-be1e01adb4e9","Type":"ContainerStarted","Data":"106fadccc1e173cbd6cdd9f144155054860f5f92762adcee53c7599c99f4361e"} Dec 10 09:55:21 crc kubenswrapper[4715]: I1210 09:55:21.398577 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 09:55:21 crc kubenswrapper[4715]: I1210 09:55:21.411307 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" Dec 10 09:55:21 crc kubenswrapper[4715]: I1210 09:55:21.411972 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-q6kwd" event={"ID":"7742e050-7183-4c20-9367-618075abadc8","Type":"ContainerDied","Data":"ab98d930b1a679fce407e8a91fdd6197361eec9fc3f11fcbe8bd5b6ca3cd6b44"} Dec 10 09:55:21 crc kubenswrapper[4715]: I1210 09:55:21.412164 4715 scope.go:117] "RemoveContainer" containerID="776f87f56130cb887230b8f13542b66aed7e25ffb3cb49306209bd93f96e9ca7" Dec 10 09:55:21 crc kubenswrapper[4715]: I1210 09:55:21.442369 4715 scope.go:117] "RemoveContainer" containerID="e5f46352a854686008a562fc489d85290d8a61903fa5abeb4653f35219f49bd2" Dec 10 09:55:21 crc kubenswrapper[4715]: I1210 09:55:21.451285 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.979862669 podStartE2EDuration="5.4512637s" podCreationTimestamp="2025-12-10 09:55:16 +0000 UTC" firstStartedPulling="2025-12-10 09:55:17.466623049 +0000 UTC m=+1280.210169300" lastFinishedPulling="2025-12-10 09:55:20.93802408 +0000 UTC m=+1283.681570331" observedRunningTime="2025-12-10 09:55:21.42322273 +0000 UTC m=+1284.166769001" watchObservedRunningTime="2025-12-10 09:55:21.4512637 +0000 UTC m=+1284.194809951" Dec 10 09:55:21 crc kubenswrapper[4715]: I1210 09:55:21.486194 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q6kwd"] Dec 10 09:55:21 crc kubenswrapper[4715]: I1210 09:55:21.495177 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q6kwd"] Dec 10 09:55:21 crc kubenswrapper[4715]: I1210 09:55:21.615095 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7742e050-7183-4c20-9367-618075abadc8" path="/var/lib/kubelet/pods/7742e050-7183-4c20-9367-618075abadc8/volumes" Dec 10 09:55:25 crc kubenswrapper[4715]: I1210 09:55:25.451886 4715 generic.go:334] "Generic (PLEG): container finished" podID="f07f968f-58f0-46c6-ac4e-d0cdcb966d10" containerID="69f349fcb33c1d2a5a8883c637c243d58b1b07882c0e4513dd824035be60d472" exitCode=0 Dec 10 09:55:25 crc kubenswrapper[4715]: I1210 09:55:25.452383 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xgcdc" event={"ID":"f07f968f-58f0-46c6-ac4e-d0cdcb966d10","Type":"ContainerDied","Data":"69f349fcb33c1d2a5a8883c637c243d58b1b07882c0e4513dd824035be60d472"} Dec 10 09:55:26 crc kubenswrapper[4715]: I1210 09:55:26.724352 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 09:55:26 crc kubenswrapper[4715]: I1210 09:55:26.724716 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 09:55:26 crc kubenswrapper[4715]: I1210 09:55:26.879534 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.039596 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-combined-ca-bundle\") pod \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.039697 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6mpq\" (UniqueName: \"kubernetes.io/projected/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-kube-api-access-p6mpq\") pod \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.039725 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-config-data\") pod \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.039863 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-scripts\") pod \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\" (UID: \"f07f968f-58f0-46c6-ac4e-d0cdcb966d10\") " Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.046230 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-scripts" (OuterVolumeSpecName: "scripts") pod "f07f968f-58f0-46c6-ac4e-d0cdcb966d10" (UID: "f07f968f-58f0-46c6-ac4e-d0cdcb966d10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.055979 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-kube-api-access-p6mpq" (OuterVolumeSpecName: "kube-api-access-p6mpq") pod "f07f968f-58f0-46c6-ac4e-d0cdcb966d10" (UID: "f07f968f-58f0-46c6-ac4e-d0cdcb966d10"). InnerVolumeSpecName "kube-api-access-p6mpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.072998 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-config-data" (OuterVolumeSpecName: "config-data") pod "f07f968f-58f0-46c6-ac4e-d0cdcb966d10" (UID: "f07f968f-58f0-46c6-ac4e-d0cdcb966d10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.075511 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f07f968f-58f0-46c6-ac4e-d0cdcb966d10" (UID: "f07f968f-58f0-46c6-ac4e-d0cdcb966d10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.141901 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.141981 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.141994 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6mpq\" (UniqueName: \"kubernetes.io/projected/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-kube-api-access-p6mpq\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.142003 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f07f968f-58f0-46c6-ac4e-d0cdcb966d10-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.476715 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xgcdc" event={"ID":"f07f968f-58f0-46c6-ac4e-d0cdcb966d10","Type":"ContainerDied","Data":"63e2bd24dfc5b44396dd9a08390b0b35fe49719a734ba2b5338f164d370f01d1"} Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.477157 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63e2bd24dfc5b44396dd9a08390b0b35fe49719a734ba2b5338f164d370f01d1" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.477110 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xgcdc" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.685991 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.686342 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerName="nova-api-log" containerID="cri-o://504cbd12ed32a575d14a2d5f2c18bd1986a8723d9f51b3718de51edef6c82de9" gracePeriod=30 Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.686499 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerName="nova-api-api" containerID="cri-o://e58beba0294f8367482faeb2cd43e629cd66839116f13421d18b41aa9e4b5a2d" gracePeriod=30 Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.703176 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": EOF" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.703383 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": EOF" Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.710395 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.710781 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="13438f95-4e2b-4225-afa0-0b2c13a9b4b2" containerName="nova-scheduler-scheduler" containerID="cri-o://53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80" gracePeriod=30 Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.723393 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.723656 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-log" containerID="cri-o://d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5" gracePeriod=30 Dec 10 09:55:27 crc kubenswrapper[4715]: I1210 09:55:27.724170 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-metadata" containerID="cri-o://9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d" gracePeriod=30 Dec 10 09:55:28 crc kubenswrapper[4715]: I1210 09:55:28.488933 4715 generic.go:334] "Generic (PLEG): container finished" podID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerID="d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5" exitCode=143 Dec 10 09:55:28 crc kubenswrapper[4715]: I1210 09:55:28.489026 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f","Type":"ContainerDied","Data":"d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5"} Dec 10 09:55:28 crc kubenswrapper[4715]: I1210 09:55:28.492079 4715 generic.go:334] "Generic (PLEG): container finished" podID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerID="504cbd12ed32a575d14a2d5f2c18bd1986a8723d9f51b3718de51edef6c82de9" exitCode=143 Dec 10 09:55:28 crc kubenswrapper[4715]: I1210 09:55:28.492135 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c","Type":"ContainerDied","Data":"504cbd12ed32a575d14a2d5f2c18bd1986a8723d9f51b3718de51edef6c82de9"} Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.350549 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.450368 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-combined-ca-bundle\") pod \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.511211 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" (UID: "378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.534844 4715 generic.go:334] "Generic (PLEG): container finished" podID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerID="9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d" exitCode=0 Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.534945 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f","Type":"ContainerDied","Data":"9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d"} Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.534980 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f","Type":"ContainerDied","Data":"c788d1cd2c1f3e981aa5344f12667b5ec607e6cfd81fa9973c2556c5281242b7"} Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.535008 4715 scope.go:117] "RemoveContainer" containerID="9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.535236 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.552672 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-logs\") pod \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.552762 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-config-data\") pod \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.552797 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-899vz\" (UniqueName: \"kubernetes.io/projected/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-kube-api-access-899vz\") pod \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.552858 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-nova-metadata-tls-certs\") pod \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\" (UID: \"378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f\") " Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.553339 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.559031 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-logs" (OuterVolumeSpecName: "logs") pod "378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" (UID: "378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.569392 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-kube-api-access-899vz" (OuterVolumeSpecName: "kube-api-access-899vz") pod "378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" (UID: "378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f"). InnerVolumeSpecName "kube-api-access-899vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.588184 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-config-data" (OuterVolumeSpecName: "config-data") pod "378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" (UID: "378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.620733 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" (UID: "378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.626982 4715 scope.go:117] "RemoveContainer" containerID="d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.655997 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.656037 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.656054 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-899vz\" (UniqueName: \"kubernetes.io/projected/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-kube-api-access-899vz\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.656064 4715 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.718740 4715 scope.go:117] "RemoveContainer" containerID="9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d" Dec 10 09:55:31 crc kubenswrapper[4715]: E1210 09:55:31.719443 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d\": container with ID starting with 9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d not found: ID does not exist" containerID="9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.719478 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d"} err="failed to get container status \"9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d\": rpc error: code = NotFound desc = could not find container \"9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d\": container with ID starting with 9625687d810a2929d66fba01b4f097c740c7605575557f62d187402b441cbf4d not found: ID does not exist" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.719501 4715 scope.go:117] "RemoveContainer" containerID="d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5" Dec 10 09:55:31 crc kubenswrapper[4715]: E1210 09:55:31.719775 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5\": container with ID starting with d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5 not found: ID does not exist" containerID="d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.719805 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5"} err="failed to get container status \"d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5\": rpc error: code = NotFound desc = could not find container \"d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5\": container with ID starting with d5bf49b5577ed7e491fac6043bfd8da5063604f3e57e17d3de45c37587766ce5 not found: ID does not exist" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.868075 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.884047 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.891449 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:55:31 crc kubenswrapper[4715]: E1210 09:55:31.892011 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-metadata" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.892028 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-metadata" Dec 10 09:55:31 crc kubenswrapper[4715]: E1210 09:55:31.892064 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7742e050-7183-4c20-9367-618075abadc8" containerName="init" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.892070 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7742e050-7183-4c20-9367-618075abadc8" containerName="init" Dec 10 09:55:31 crc kubenswrapper[4715]: E1210 09:55:31.892087 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f07f968f-58f0-46c6-ac4e-d0cdcb966d10" containerName="nova-manage" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.892092 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f07f968f-58f0-46c6-ac4e-d0cdcb966d10" containerName="nova-manage" Dec 10 09:55:31 crc kubenswrapper[4715]: E1210 09:55:31.892105 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-log" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.892112 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-log" Dec 10 09:55:31 crc kubenswrapper[4715]: E1210 09:55:31.892127 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7742e050-7183-4c20-9367-618075abadc8" containerName="dnsmasq-dns" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.892132 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7742e050-7183-4c20-9367-618075abadc8" containerName="dnsmasq-dns" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.906804 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-log" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.906842 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7742e050-7183-4c20-9367-618075abadc8" containerName="dnsmasq-dns" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.906853 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-metadata" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.906868 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f07f968f-58f0-46c6-ac4e-d0cdcb966d10" containerName="nova-manage" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.908893 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.909561 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.911142 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.911336 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.911432 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.964004 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-combined-ca-bundle\") pod \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.964112 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f9qt\" (UniqueName: \"kubernetes.io/projected/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-kube-api-access-9f9qt\") pod \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.964234 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-config-data\") pod \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\" (UID: \"13438f95-4e2b-4225-afa0-0b2c13a9b4b2\") " Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.964873 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-logs\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.964927 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.964956 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-config-data\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.964986 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ldcj\" (UniqueName: \"kubernetes.io/projected/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-kube-api-access-7ldcj\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.965027 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.975289 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-kube-api-access-9f9qt" (OuterVolumeSpecName: "kube-api-access-9f9qt") pod "13438f95-4e2b-4225-afa0-0b2c13a9b4b2" (UID: "13438f95-4e2b-4225-afa0-0b2c13a9b4b2"). InnerVolumeSpecName "kube-api-access-9f9qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.991297 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13438f95-4e2b-4225-afa0-0b2c13a9b4b2" (UID: "13438f95-4e2b-4225-afa0-0b2c13a9b4b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:31 crc kubenswrapper[4715]: I1210 09:55:31.993229 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-config-data" (OuterVolumeSpecName: "config-data") pod "13438f95-4e2b-4225-afa0-0b2c13a9b4b2" (UID: "13438f95-4e2b-4225-afa0-0b2c13a9b4b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.067513 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-logs\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.067587 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.067622 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-config-data\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.067668 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ldcj\" (UniqueName: \"kubernetes.io/projected/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-kube-api-access-7ldcj\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.067719 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.067884 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f9qt\" (UniqueName: \"kubernetes.io/projected/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-kube-api-access-9f9qt\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.067924 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.067939 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13438f95-4e2b-4225-afa0-0b2c13a9b4b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.068103 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-logs\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.070987 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.071389 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.072100 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-config-data\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.088423 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ldcj\" (UniqueName: \"kubernetes.io/projected/8027e8a4-9be7-4764-8c1d-f25a3eb3587b-kube-api-access-7ldcj\") pod \"nova-metadata-0\" (UID: \"8027e8a4-9be7-4764-8c1d-f25a3eb3587b\") " pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.234956 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.579573 4715 generic.go:334] "Generic (PLEG): container finished" podID="13438f95-4e2b-4225-afa0-0b2c13a9b4b2" containerID="53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80" exitCode=0 Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.579944 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"13438f95-4e2b-4225-afa0-0b2c13a9b4b2","Type":"ContainerDied","Data":"53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80"} Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.579994 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"13438f95-4e2b-4225-afa0-0b2c13a9b4b2","Type":"ContainerDied","Data":"1600f912e42970dab0543e5746c65164a07604fc45e05ba5cba1dd8b91015134"} Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.580014 4715 scope.go:117] "RemoveContainer" containerID="53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.580120 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.603129 4715 scope.go:117] "RemoveContainer" containerID="53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80" Dec 10 09:55:32 crc kubenswrapper[4715]: E1210 09:55:32.603846 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80\": container with ID starting with 53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80 not found: ID does not exist" containerID="53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.604287 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80"} err="failed to get container status \"53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80\": rpc error: code = NotFound desc = could not find container \"53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80\": container with ID starting with 53a60e3f5688ae0a9d514ecc75388881dc4173ba79229a9046849b98651bee80 not found: ID does not exist" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.620110 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.658786 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.668943 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:55:32 crc kubenswrapper[4715]: E1210 09:55:32.670558 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13438f95-4e2b-4225-afa0-0b2c13a9b4b2" containerName="nova-scheduler-scheduler" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.670579 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="13438f95-4e2b-4225-afa0-0b2c13a9b4b2" containerName="nova-scheduler-scheduler" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.670803 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="13438f95-4e2b-4225-afa0-0b2c13a9b4b2" containerName="nova-scheduler-scheduler" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.671517 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.676619 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.690446 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:55:32 crc kubenswrapper[4715]: W1210 09:55:32.729000 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8027e8a4_9be7_4764_8c1d_f25a3eb3587b.slice/crio-564bfae1d86e25cdc262f4681dbc1665febf7d416bbed15b50e93879d40d454b WatchSource:0}: Error finding container 564bfae1d86e25cdc262f4681dbc1665febf7d416bbed15b50e93879d40d454b: Status 404 returned error can't find the container with id 564bfae1d86e25cdc262f4681dbc1665febf7d416bbed15b50e93879d40d454b Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.731161 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.781641 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr2gh\" (UniqueName: \"kubernetes.io/projected/6feb8a72-37ea-419b-b69e-0a2ef81d79ca-kube-api-access-kr2gh\") pod \"nova-scheduler-0\" (UID: \"6feb8a72-37ea-419b-b69e-0a2ef81d79ca\") " pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.781691 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6feb8a72-37ea-419b-b69e-0a2ef81d79ca-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6feb8a72-37ea-419b-b69e-0a2ef81d79ca\") " pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.781833 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6feb8a72-37ea-419b-b69e-0a2ef81d79ca-config-data\") pod \"nova-scheduler-0\" (UID: \"6feb8a72-37ea-419b-b69e-0a2ef81d79ca\") " pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.883697 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr2gh\" (UniqueName: \"kubernetes.io/projected/6feb8a72-37ea-419b-b69e-0a2ef81d79ca-kube-api-access-kr2gh\") pod \"nova-scheduler-0\" (UID: \"6feb8a72-37ea-419b-b69e-0a2ef81d79ca\") " pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.883756 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6feb8a72-37ea-419b-b69e-0a2ef81d79ca-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6feb8a72-37ea-419b-b69e-0a2ef81d79ca\") " pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.883879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6feb8a72-37ea-419b-b69e-0a2ef81d79ca-config-data\") pod \"nova-scheduler-0\" (UID: \"6feb8a72-37ea-419b-b69e-0a2ef81d79ca\") " pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.890721 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6feb8a72-37ea-419b-b69e-0a2ef81d79ca-config-data\") pod \"nova-scheduler-0\" (UID: \"6feb8a72-37ea-419b-b69e-0a2ef81d79ca\") " pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.891212 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6feb8a72-37ea-419b-b69e-0a2ef81d79ca-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6feb8a72-37ea-419b-b69e-0a2ef81d79ca\") " pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.901147 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr2gh\" (UniqueName: \"kubernetes.io/projected/6feb8a72-37ea-419b-b69e-0a2ef81d79ca-kube-api-access-kr2gh\") pod \"nova-scheduler-0\" (UID: \"6feb8a72-37ea-419b-b69e-0a2ef81d79ca\") " pod="openstack/nova-scheduler-0" Dec 10 09:55:32 crc kubenswrapper[4715]: I1210 09:55:32.987273 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.461622 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.602185 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8027e8a4-9be7-4764-8c1d-f25a3eb3587b","Type":"ContainerStarted","Data":"4223ff36c3800e28938a7e8a63268f53cf6219322c7ef920323874ed27114277"} Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.602589 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8027e8a4-9be7-4764-8c1d-f25a3eb3587b","Type":"ContainerStarted","Data":"225fed10b952e5439ab90736e6bb14a31220f87fe90ab844a28cc2b78eca8180"} Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.602606 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8027e8a4-9be7-4764-8c1d-f25a3eb3587b","Type":"ContainerStarted","Data":"564bfae1d86e25cdc262f4681dbc1665febf7d416bbed15b50e93879d40d454b"} Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.612347 4715 generic.go:334] "Generic (PLEG): container finished" podID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerID="e58beba0294f8367482faeb2cd43e629cd66839116f13421d18b41aa9e4b5a2d" exitCode=0 Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.622258 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13438f95-4e2b-4225-afa0-0b2c13a9b4b2" path="/var/lib/kubelet/pods/13438f95-4e2b-4225-afa0-0b2c13a9b4b2/volumes" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.628482 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.628460398 podStartE2EDuration="2.628460398s" podCreationTimestamp="2025-12-10 09:55:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:55:33.621373792 +0000 UTC m=+1296.364920063" watchObservedRunningTime="2025-12-10 09:55:33.628460398 +0000 UTC m=+1296.372006649" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.628736 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" path="/var/lib/kubelet/pods/378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f/volumes" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.629958 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6feb8a72-37ea-419b-b69e-0a2ef81d79ca","Type":"ContainerStarted","Data":"66845fe6d08aa17b073617ff283351b744d909847334318e0af5fcb89b0750c0"} Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.629997 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c","Type":"ContainerDied","Data":"e58beba0294f8367482faeb2cd43e629cd66839116f13421d18b41aa9e4b5a2d"} Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.659881 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.823562 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-public-tls-certs\") pod \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.823708 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-config-data\") pod \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.823800 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-internal-tls-certs\") pod \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.823834 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-logs\") pod \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.823889 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7sqx\" (UniqueName: \"kubernetes.io/projected/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-kube-api-access-q7sqx\") pod \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.823936 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-combined-ca-bundle\") pod \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\" (UID: \"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c\") " Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.824825 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-logs" (OuterVolumeSpecName: "logs") pod "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" (UID: "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.833129 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-kube-api-access-q7sqx" (OuterVolumeSpecName: "kube-api-access-q7sqx") pod "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" (UID: "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c"). InnerVolumeSpecName "kube-api-access-q7sqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.855416 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-config-data" (OuterVolumeSpecName: "config-data") pod "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" (UID: "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.857081 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" (UID: "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.875816 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" (UID: "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.894712 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" (UID: "c443ab3b-2046-4567-ab9b-0cbc0cf79e0c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.926505 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.926549 4715 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.926567 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-logs\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.926581 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7sqx\" (UniqueName: \"kubernetes.io/projected/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-kube-api-access-q7sqx\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.926595 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:33 crc kubenswrapper[4715]: I1210 09:55:33.926606 4715 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.625157 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6feb8a72-37ea-419b-b69e-0a2ef81d79ca","Type":"ContainerStarted","Data":"afdf049b3b9c172e34ba2f79825300313923eee9b65a24a24f58b91c0d31b347"} Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.628823 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.629107 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c443ab3b-2046-4567-ab9b-0cbc0cf79e0c","Type":"ContainerDied","Data":"69b2c45d467bd786fa882f5cf842583639f3ec55fdf69e33879f5b4768f81cdd"} Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.629155 4715 scope.go:117] "RemoveContainer" containerID="e58beba0294f8367482faeb2cd43e629cd66839116f13421d18b41aa9e4b5a2d" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.644745 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.644726981 podStartE2EDuration="2.644726981s" podCreationTimestamp="2025-12-10 09:55:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:55:34.644590577 +0000 UTC m=+1297.388136828" watchObservedRunningTime="2025-12-10 09:55:34.644726981 +0000 UTC m=+1297.388273232" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.658205 4715 scope.go:117] "RemoveContainer" containerID="504cbd12ed32a575d14a2d5f2c18bd1986a8723d9f51b3718de51edef6c82de9" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.677170 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.691500 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.706011 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:34 crc kubenswrapper[4715]: E1210 09:55:34.706481 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerName="nova-api-api" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.706493 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerName="nova-api-api" Dec 10 09:55:34 crc kubenswrapper[4715]: E1210 09:55:34.706515 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerName="nova-api-log" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.706521 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerName="nova-api-log" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.706697 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerName="nova-api-log" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.706710 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" containerName="nova-api-api" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.707850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.711179 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.711366 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.711571 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.722328 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.844526 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22d2k\" (UniqueName: \"kubernetes.io/projected/2648041d-c409-4012-b90b-3d780bacb44d-kube-api-access-22d2k\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.844669 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-config-data\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.844743 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.844830 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-public-tls-certs\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.844879 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.844947 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2648041d-c409-4012-b90b-3d780bacb44d-logs\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.946686 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22d2k\" (UniqueName: \"kubernetes.io/projected/2648041d-c409-4012-b90b-3d780bacb44d-kube-api-access-22d2k\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.947104 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-config-data\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.947263 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.947311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-public-tls-certs\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.947338 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.947375 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2648041d-c409-4012-b90b-3d780bacb44d-logs\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.947765 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2648041d-c409-4012-b90b-3d780bacb44d-logs\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.953262 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.953269 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.954725 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-config-data\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.960990 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2648041d-c409-4012-b90b-3d780bacb44d-public-tls-certs\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:34 crc kubenswrapper[4715]: I1210 09:55:34.963593 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22d2k\" (UniqueName: \"kubernetes.io/projected/2648041d-c409-4012-b90b-3d780bacb44d-kube-api-access-22d2k\") pod \"nova-api-0\" (UID: \"2648041d-c409-4012-b90b-3d780bacb44d\") " pod="openstack/nova-api-0" Dec 10 09:55:35 crc kubenswrapper[4715]: I1210 09:55:35.029795 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 09:55:35 crc kubenswrapper[4715]: I1210 09:55:35.520236 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 09:55:35 crc kubenswrapper[4715]: I1210 09:55:35.635111 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c443ab3b-2046-4567-ab9b-0cbc0cf79e0c" path="/var/lib/kubelet/pods/c443ab3b-2046-4567-ab9b-0cbc0cf79e0c/volumes" Dec 10 09:55:35 crc kubenswrapper[4715]: I1210 09:55:35.641703 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2648041d-c409-4012-b90b-3d780bacb44d","Type":"ContainerStarted","Data":"59e7627a5a22d33aaf2cb34494847b9bcbe28d27ae9773b4c16a1789c08dd54d"} Dec 10 09:55:36 crc kubenswrapper[4715]: I1210 09:55:36.284037 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": dial tcp 10.217.0.193:8775: i/o timeout" Dec 10 09:55:36 crc kubenswrapper[4715]: I1210 09:55:36.284214 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="378f8c2b-6bd7-4440-8a6c-49c76ebe9d0f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": dial tcp 10.217.0.193:8775: i/o timeout" Dec 10 09:55:36 crc kubenswrapper[4715]: I1210 09:55:36.663302 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2648041d-c409-4012-b90b-3d780bacb44d","Type":"ContainerStarted","Data":"877e4e1a244446b6e71c945318fa9c5859f2f543f8a8b447f21ec66d42601ffe"} Dec 10 09:55:36 crc kubenswrapper[4715]: I1210 09:55:36.663386 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2648041d-c409-4012-b90b-3d780bacb44d","Type":"ContainerStarted","Data":"42c18a4c00007323cefb30e8d81e22ab5e54e43bf5d1aaf8e674bb3d65d20024"} Dec 10 09:55:37 crc kubenswrapper[4715]: I1210 09:55:37.235867 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 09:55:37 crc kubenswrapper[4715]: I1210 09:55:37.235996 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 09:55:37 crc kubenswrapper[4715]: I1210 09:55:37.988474 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 09:55:42 crc kubenswrapper[4715]: I1210 09:55:42.236008 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 09:55:42 crc kubenswrapper[4715]: I1210 09:55:42.236425 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 09:55:42 crc kubenswrapper[4715]: I1210 09:55:42.988589 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 09:55:43 crc kubenswrapper[4715]: I1210 09:55:43.021485 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 09:55:43 crc kubenswrapper[4715]: I1210 09:55:43.053818 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=9.053795786 podStartE2EDuration="9.053795786s" podCreationTimestamp="2025-12-10 09:55:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:55:36.685032353 +0000 UTC m=+1299.428578604" watchObservedRunningTime="2025-12-10 09:55:43.053795786 +0000 UTC m=+1305.797342037" Dec 10 09:55:43 crc kubenswrapper[4715]: I1210 09:55:43.254302 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8027e8a4-9be7-4764-8c1d-f25a3eb3587b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 09:55:43 crc kubenswrapper[4715]: I1210 09:55:43.254315 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8027e8a4-9be7-4764-8c1d-f25a3eb3587b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 09:55:43 crc kubenswrapper[4715]: I1210 09:55:43.951845 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 09:55:45 crc kubenswrapper[4715]: I1210 09:55:45.031141 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 09:55:45 crc kubenswrapper[4715]: I1210 09:55:45.031205 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 09:55:46 crc kubenswrapper[4715]: I1210 09:55:46.045090 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2648041d-c409-4012-b90b-3d780bacb44d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 09:55:46 crc kubenswrapper[4715]: I1210 09:55:46.045111 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2648041d-c409-4012-b90b-3d780bacb44d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 09:55:46 crc kubenswrapper[4715]: I1210 09:55:46.976777 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 09:55:52 crc kubenswrapper[4715]: I1210 09:55:52.245446 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 09:55:52 crc kubenswrapper[4715]: I1210 09:55:52.246367 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 09:55:52 crc kubenswrapper[4715]: I1210 09:55:52.254269 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 09:55:53 crc kubenswrapper[4715]: I1210 09:55:53.064595 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 09:55:55 crc kubenswrapper[4715]: I1210 09:55:55.038755 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 09:55:55 crc kubenswrapper[4715]: I1210 09:55:55.040772 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 09:55:55 crc kubenswrapper[4715]: I1210 09:55:55.040939 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 09:55:55 crc kubenswrapper[4715]: I1210 09:55:55.052577 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 09:55:55 crc kubenswrapper[4715]: I1210 09:55:55.248102 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 09:55:55 crc kubenswrapper[4715]: I1210 09:55:55.254468 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 09:56:04 crc kubenswrapper[4715]: I1210 09:56:04.035343 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 09:56:04 crc kubenswrapper[4715]: I1210 09:56:04.887731 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 09:56:08 crc kubenswrapper[4715]: I1210 09:56:08.358937 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="ad27e35f-aceb-42a8-a60f-efd71d0a3b73" containerName="rabbitmq" containerID="cri-o://64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653" gracePeriod=604796 Dec 10 09:56:09 crc kubenswrapper[4715]: I1210 09:56:09.263223 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="5afc401a-b98d-4b26-8a47-c18582062f49" containerName="rabbitmq" containerID="cri-o://e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b" gracePeriod=604796 Dec 10 09:56:10 crc kubenswrapper[4715]: I1210 09:56:10.374084 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="ad27e35f-aceb-42a8-a60f-efd71d0a3b73" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.96:5671: connect: connection refused" Dec 10 09:56:10 crc kubenswrapper[4715]: I1210 09:56:10.779431 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="5afc401a-b98d-4b26-8a47-c18582062f49" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.025025 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138174 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-erlang-cookie-secret\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138235 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwmg5\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-kube-api-access-jwmg5\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138296 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-config-data\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138340 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-erlang-cookie\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138385 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-confd\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138416 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138465 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-tls\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138496 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-pod-info\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138548 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-plugins-conf\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138605 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-server-conf\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.138663 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-plugins\") pod \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\" (UID: \"ad27e35f-aceb-42a8-a60f-efd71d0a3b73\") " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.139184 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.140442 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.141665 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.148598 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.152136 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-kube-api-access-jwmg5" (OuterVolumeSpecName: "kube-api-access-jwmg5") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "kube-api-access-jwmg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.157946 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-pod-info" (OuterVolumeSpecName: "pod-info") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.164588 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.176427 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.196147 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-config-data" (OuterVolumeSpecName: "config-data") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.238324 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-server-conf" (OuterVolumeSpecName: "server-conf") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.242799 4715 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-server-conf\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.242832 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.242844 4715 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.242856 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwmg5\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-kube-api-access-jwmg5\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.242870 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.242880 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.242941 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.242956 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.242967 4715 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-pod-info\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.242977 4715 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.286589 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.294533 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ad27e35f-aceb-42a8-a60f-efd71d0a3b73" (UID: "ad27e35f-aceb-42a8-a60f-efd71d0a3b73"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.347756 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad27e35f-aceb-42a8-a60f-efd71d0a3b73-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.347816 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.457109 4715 generic.go:334] "Generic (PLEG): container finished" podID="ad27e35f-aceb-42a8-a60f-efd71d0a3b73" containerID="64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653" exitCode=0 Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.457189 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad27e35f-aceb-42a8-a60f-efd71d0a3b73","Type":"ContainerDied","Data":"64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653"} Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.457208 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.457242 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad27e35f-aceb-42a8-a60f-efd71d0a3b73","Type":"ContainerDied","Data":"d7416f7a399cf0025b65f3f516759f0a43dd20c45091f846fb5ef2446c05debb"} Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.457262 4715 scope.go:117] "RemoveContainer" containerID="64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.517540 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.535719 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.561268 4715 scope.go:117] "RemoveContainer" containerID="ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.561861 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 09:56:15 crc kubenswrapper[4715]: E1210 09:56:15.562645 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad27e35f-aceb-42a8-a60f-efd71d0a3b73" containerName="setup-container" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.562677 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad27e35f-aceb-42a8-a60f-efd71d0a3b73" containerName="setup-container" Dec 10 09:56:15 crc kubenswrapper[4715]: E1210 09:56:15.562713 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad27e35f-aceb-42a8-a60f-efd71d0a3b73" containerName="rabbitmq" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.562721 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad27e35f-aceb-42a8-a60f-efd71d0a3b73" containerName="rabbitmq" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.563027 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad27e35f-aceb-42a8-a60f-efd71d0a3b73" containerName="rabbitmq" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.565442 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.571597 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.571841 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.572147 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.572175 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-85zs9" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.572271 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.572470 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.572635 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.577508 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.640136 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad27e35f-aceb-42a8-a60f-efd71d0a3b73" path="/var/lib/kubelet/pods/ad27e35f-aceb-42a8-a60f-efd71d0a3b73/volumes" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.664525 4715 scope.go:117] "RemoveContainer" containerID="64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653" Dec 10 09:56:15 crc kubenswrapper[4715]: E1210 09:56:15.679352 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653\": container with ID starting with 64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653 not found: ID does not exist" containerID="64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.679421 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653"} err="failed to get container status \"64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653\": rpc error: code = NotFound desc = could not find container \"64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653\": container with ID starting with 64679f529eaf14f514250a5eeccc8c682eb9639e5b3824fbebfeccd76e1b4653 not found: ID does not exist" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.679457 4715 scope.go:117] "RemoveContainer" containerID="ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2" Dec 10 09:56:15 crc kubenswrapper[4715]: E1210 09:56:15.680296 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2\": container with ID starting with ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2 not found: ID does not exist" containerID="ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.680349 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2"} err="failed to get container status \"ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2\": rpc error: code = NotFound desc = could not find container \"ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2\": container with ID starting with ef9c3ae88a71b6aefb9cabf4057ffa86f6b8d085fd0872094a0f78c83e98a0d2 not found: ID does not exist" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756468 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4e8933c-d768-48cb-bac7-245abd8761ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756572 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkhfv\" (UniqueName: \"kubernetes.io/projected/e4e8933c-d768-48cb-bac7-245abd8761ad-kube-api-access-vkhfv\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756618 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4e8933c-d768-48cb-bac7-245abd8761ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756661 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4e8933c-d768-48cb-bac7-245abd8761ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756743 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756770 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756799 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756832 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756853 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756877 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4e8933c-d768-48cb-bac7-245abd8761ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.756937 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4e8933c-d768-48cb-bac7-245abd8761ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.859396 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4e8933c-d768-48cb-bac7-245abd8761ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.859810 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkhfv\" (UniqueName: \"kubernetes.io/projected/e4e8933c-d768-48cb-bac7-245abd8761ad-kube-api-access-vkhfv\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.860227 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4e8933c-d768-48cb-bac7-245abd8761ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.860286 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4e8933c-d768-48cb-bac7-245abd8761ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.860376 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.860412 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.860472 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.860508 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.860535 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.860559 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4e8933c-d768-48cb-bac7-245abd8761ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.860603 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4e8933c-d768-48cb-bac7-245abd8761ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.861298 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4e8933c-d768-48cb-bac7-245abd8761ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.861412 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.862098 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4e8933c-d768-48cb-bac7-245abd8761ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.862364 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.862868 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4e8933c-d768-48cb-bac7-245abd8761ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.865908 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4e8933c-d768-48cb-bac7-245abd8761ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.866985 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.868796 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.872691 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4e8933c-d768-48cb-bac7-245abd8761ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.874966 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4e8933c-d768-48cb-bac7-245abd8761ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.883764 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkhfv\" (UniqueName: \"kubernetes.io/projected/e4e8933c-d768-48cb-bac7-245abd8761ad-kube-api-access-vkhfv\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.907242 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"e4e8933c-d768-48cb-bac7-245abd8761ad\") " pod="openstack/rabbitmq-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.976534 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:15 crc kubenswrapper[4715]: I1210 09:56:15.991096 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.067716 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5afc401a-b98d-4b26-8a47-c18582062f49-erlang-cookie-secret\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.067844 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-tls\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.067941 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5afc401a-b98d-4b26-8a47-c18582062f49-pod-info\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.068064 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-config-data\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.068110 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-plugins-conf\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.068147 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-plugins\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.068174 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-server-conf\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.068218 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l2mv\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-kube-api-access-4l2mv\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.068246 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.068294 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-erlang-cookie\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.068321 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-confd\") pod \"5afc401a-b98d-4b26-8a47-c18582062f49\" (UID: \"5afc401a-b98d-4b26-8a47-c18582062f49\") " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.069263 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.071988 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.072240 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.080691 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.080814 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5afc401a-b98d-4b26-8a47-c18582062f49-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.081867 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5afc401a-b98d-4b26-8a47-c18582062f49-pod-info" (OuterVolumeSpecName: "pod-info") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.084904 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.093018 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-kube-api-access-4l2mv" (OuterVolumeSpecName: "kube-api-access-4l2mv") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "kube-api-access-4l2mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.131656 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-config-data" (OuterVolumeSpecName: "config-data") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.163755 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-server-conf" (OuterVolumeSpecName: "server-conf") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.172313 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.172381 4715 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.172402 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.172413 4715 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5afc401a-b98d-4b26-8a47-c18582062f49-server-conf\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.172430 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l2mv\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-kube-api-access-4l2mv\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.172489 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.172514 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.172530 4715 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5afc401a-b98d-4b26-8a47-c18582062f49-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.172539 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.172558 4715 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5afc401a-b98d-4b26-8a47-c18582062f49-pod-info\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.199063 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.217465 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5afc401a-b98d-4b26-8a47-c18582062f49" (UID: "5afc401a-b98d-4b26-8a47-c18582062f49"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.274297 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.274340 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5afc401a-b98d-4b26-8a47-c18582062f49-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.473257 4715 generic.go:334] "Generic (PLEG): container finished" podID="5afc401a-b98d-4b26-8a47-c18582062f49" containerID="e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b" exitCode=0 Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.473310 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5afc401a-b98d-4b26-8a47-c18582062f49","Type":"ContainerDied","Data":"e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b"} Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.473361 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5afc401a-b98d-4b26-8a47-c18582062f49","Type":"ContainerDied","Data":"4a81c8b991df4ee7ce5500dc6e43bf36260001444db6237d0ca7fe62778e813c"} Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.473385 4715 scope.go:117] "RemoveContainer" containerID="e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.473413 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.517287 4715 scope.go:117] "RemoveContainer" containerID="993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.553827 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.571328 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.574951 4715 scope.go:117] "RemoveContainer" containerID="e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b" Dec 10 09:56:16 crc kubenswrapper[4715]: E1210 09:56:16.575531 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b\": container with ID starting with e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b not found: ID does not exist" containerID="e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.575597 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b"} err="failed to get container status \"e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b\": rpc error: code = NotFound desc = could not find container \"e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b\": container with ID starting with e68615c07bfdc1f86a967462639d249c3e463f3911e0e256112270704a41d32b not found: ID does not exist" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.575633 4715 scope.go:117] "RemoveContainer" containerID="993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c" Dec 10 09:56:16 crc kubenswrapper[4715]: E1210 09:56:16.576365 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c\": container with ID starting with 993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c not found: ID does not exist" containerID="993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.576397 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c"} err="failed to get container status \"993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c\": rpc error: code = NotFound desc = could not find container \"993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c\": container with ID starting with 993a2efa33084df9ae6c45ed33137e7ceeb30b531560c2bf0e3674751544c16c not found: ID does not exist" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.580941 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 09:56:16 crc kubenswrapper[4715]: E1210 09:56:16.581531 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5afc401a-b98d-4b26-8a47-c18582062f49" containerName="setup-container" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.581556 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5afc401a-b98d-4b26-8a47-c18582062f49" containerName="setup-container" Dec 10 09:56:16 crc kubenswrapper[4715]: E1210 09:56:16.581611 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5afc401a-b98d-4b26-8a47-c18582062f49" containerName="rabbitmq" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.581621 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5afc401a-b98d-4b26-8a47-c18582062f49" containerName="rabbitmq" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.581936 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5afc401a-b98d-4b26-8a47-c18582062f49" containerName="rabbitmq" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.583410 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.589346 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.589612 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.589860 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.590119 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.590661 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.591027 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.591056 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wrvwh" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.607011 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.626777 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.685791 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf6962bf-0a77-4285-b58f-6213688e74b7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.686336 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.686373 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.686407 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.686433 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.686612 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf6962bf-0a77-4285-b58f-6213688e74b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.686686 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q87m4\" (UniqueName: \"kubernetes.io/projected/bf6962bf-0a77-4285-b58f-6213688e74b7-kube-api-access-q87m4\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.686722 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf6962bf-0a77-4285-b58f-6213688e74b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.686848 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.686872 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf6962bf-0a77-4285-b58f-6213688e74b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.686940 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf6962bf-0a77-4285-b58f-6213688e74b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789120 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789178 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf6962bf-0a77-4285-b58f-6213688e74b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789207 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf6962bf-0a77-4285-b58f-6213688e74b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789296 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf6962bf-0a77-4285-b58f-6213688e74b7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789320 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789347 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789376 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789430 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789479 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf6962bf-0a77-4285-b58f-6213688e74b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789507 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q87m4\" (UniqueName: \"kubernetes.io/projected/bf6962bf-0a77-4285-b58f-6213688e74b7-kube-api-access-q87m4\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789540 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf6962bf-0a77-4285-b58f-6213688e74b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.789766 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.790136 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.790810 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf6962bf-0a77-4285-b58f-6213688e74b7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.790819 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.791063 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf6962bf-0a77-4285-b58f-6213688e74b7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.791202 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf6962bf-0a77-4285-b58f-6213688e74b7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.794834 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf6962bf-0a77-4285-b58f-6213688e74b7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.794897 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.797519 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf6962bf-0a77-4285-b58f-6213688e74b7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.797656 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bf6962bf-0a77-4285-b58f-6213688e74b7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.829703 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q87m4\" (UniqueName: \"kubernetes.io/projected/bf6962bf-0a77-4285-b58f-6213688e74b7-kube-api-access-q87m4\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:16 crc kubenswrapper[4715]: I1210 09:56:16.847795 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf6962bf-0a77-4285-b58f-6213688e74b7\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:17 crc kubenswrapper[4715]: I1210 09:56:17.087439 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:56:17 crc kubenswrapper[4715]: I1210 09:56:17.484529 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4e8933c-d768-48cb-bac7-245abd8761ad","Type":"ContainerStarted","Data":"5f2bb47c8c93c9bc79cc42cfb940f31ba7a33d04ae4d181b1ea1b1f9f793bae7"} Dec 10 09:56:17 crc kubenswrapper[4715]: I1210 09:56:17.532900 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 09:56:17 crc kubenswrapper[4715]: W1210 09:56:17.534167 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf6962bf_0a77_4285_b58f_6213688e74b7.slice/crio-64de894e19f6fd024e99e734d5180c11145029656bec5507dea36f47ab7638f0 WatchSource:0}: Error finding container 64de894e19f6fd024e99e734d5180c11145029656bec5507dea36f47ab7638f0: Status 404 returned error can't find the container with id 64de894e19f6fd024e99e734d5180c11145029656bec5507dea36f47ab7638f0 Dec 10 09:56:17 crc kubenswrapper[4715]: I1210 09:56:17.621502 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5afc401a-b98d-4b26-8a47-c18582062f49" path="/var/lib/kubelet/pods/5afc401a-b98d-4b26-8a47-c18582062f49/volumes" Dec 10 09:56:18 crc kubenswrapper[4715]: I1210 09:56:18.495985 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4e8933c-d768-48cb-bac7-245abd8761ad","Type":"ContainerStarted","Data":"185ac203d8b588fc81ceeecce1f460e5ca65ffe6f0faa8d57781e308d4fe5c7c"} Dec 10 09:56:18 crc kubenswrapper[4715]: I1210 09:56:18.497803 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf6962bf-0a77-4285-b58f-6213688e74b7","Type":"ContainerStarted","Data":"64de894e19f6fd024e99e734d5180c11145029656bec5507dea36f47ab7638f0"} Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.348628 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lj6xg"] Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.350796 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.353699 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.374770 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lj6xg"] Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.448142 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.448823 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5vjs\" (UniqueName: \"kubernetes.io/projected/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-kube-api-access-d5vjs\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.448890 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-config\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.449062 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.449156 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.449501 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.449694 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.551723 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.551801 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.551857 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.551968 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5vjs\" (UniqueName: \"kubernetes.io/projected/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-kube-api-access-d5vjs\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.551997 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-config\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.552031 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.552048 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.553469 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-config\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.553484 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.553487 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.554051 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.554149 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.554367 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.578559 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5vjs\" (UniqueName: \"kubernetes.io/projected/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-kube-api-access-d5vjs\") pod \"dnsmasq-dns-79bd4cc8c9-lj6xg\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:19 crc kubenswrapper[4715]: I1210 09:56:19.693398 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:20 crc kubenswrapper[4715]: I1210 09:56:20.133036 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lj6xg"] Dec 10 09:56:20 crc kubenswrapper[4715]: I1210 09:56:20.518552 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf6962bf-0a77-4285-b58f-6213688e74b7","Type":"ContainerStarted","Data":"c4fabcfa33893a06e4684442f7b7a4379b662d870d5b75e0da31be85d2b2482e"} Dec 10 09:56:20 crc kubenswrapper[4715]: I1210 09:56:20.521069 4715 generic.go:334] "Generic (PLEG): container finished" podID="ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" containerID="c4b06541705dc8c65df31fb1d289723fb64c1ed04324f08f37dac45fb5bbbcb9" exitCode=0 Dec 10 09:56:20 crc kubenswrapper[4715]: I1210 09:56:20.521124 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" event={"ID":"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41","Type":"ContainerDied","Data":"c4b06541705dc8c65df31fb1d289723fb64c1ed04324f08f37dac45fb5bbbcb9"} Dec 10 09:56:20 crc kubenswrapper[4715]: I1210 09:56:20.521160 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" event={"ID":"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41","Type":"ContainerStarted","Data":"3d85f4d8c385804509745de68ffedb439fbd29781e10ebabe15a35cc0fbc8cac"} Dec 10 09:56:21 crc kubenswrapper[4715]: I1210 09:56:21.532045 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" event={"ID":"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41","Type":"ContainerStarted","Data":"5c956ce1b9205394740d68ef76851a5205f2a2d5a764ed3011ff9f1098ddafe6"} Dec 10 09:56:21 crc kubenswrapper[4715]: I1210 09:56:21.559260 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" podStartSLOduration=2.5592420750000002 podStartE2EDuration="2.559242075s" podCreationTimestamp="2025-12-10 09:56:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:56:21.553875959 +0000 UTC m=+1344.297422230" watchObservedRunningTime="2025-12-10 09:56:21.559242075 +0000 UTC m=+1344.302788326" Dec 10 09:56:22 crc kubenswrapper[4715]: I1210 09:56:22.542098 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.695080 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.753700 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fcqpx"] Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.753990 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" podUID="a7490f58-8a80-453b-924b-3f14aeceeff8" containerName="dnsmasq-dns" containerID="cri-o://031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927" gracePeriod=10 Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.840233 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" podUID="a7490f58-8a80-453b-924b-3f14aeceeff8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.198:5353: connect: connection refused" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.897056 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-g755x"] Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.899113 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.952191 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.952532 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-config\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.952579 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-dns-svc\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.952607 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d7xb\" (UniqueName: \"kubernetes.io/projected/540bd567-e127-4db2-b6c5-0d447ce7b34c-kube-api-access-6d7xb\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.952642 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.952670 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.952692 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:29 crc kubenswrapper[4715]: I1210 09:56:29.960652 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-g755x"] Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.053891 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.053997 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-config\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.054046 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-dns-svc\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.054067 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d7xb\" (UniqueName: \"kubernetes.io/projected/540bd567-e127-4db2-b6c5-0d447ce7b34c-kube-api-access-6d7xb\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.054096 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.054124 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.054143 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.054949 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.055886 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-dns-svc\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.056382 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.056486 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.056889 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.060586 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/540bd567-e127-4db2-b6c5-0d447ce7b34c-config\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.079261 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d7xb\" (UniqueName: \"kubernetes.io/projected/540bd567-e127-4db2-b6c5-0d447ce7b34c-kube-api-access-6d7xb\") pod \"dnsmasq-dns-55478c4467-g755x\" (UID: \"540bd567-e127-4db2-b6c5-0d447ce7b34c\") " pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.275085 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.472389 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.627502 4715 generic.go:334] "Generic (PLEG): container finished" podID="a7490f58-8a80-453b-924b-3f14aeceeff8" containerID="031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927" exitCode=0 Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.627549 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" event={"ID":"a7490f58-8a80-453b-924b-3f14aeceeff8","Type":"ContainerDied","Data":"031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927"} Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.627577 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" event={"ID":"a7490f58-8a80-453b-924b-3f14aeceeff8","Type":"ContainerDied","Data":"40c786517981b7605b57118882b6fef08690177aca913f5108808fb0871d09bc"} Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.627594 4715 scope.go:117] "RemoveContainer" containerID="031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.627601 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-fcqpx" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.648900 4715 scope.go:117] "RemoveContainer" containerID="5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.673889 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-config\") pod \"a7490f58-8a80-453b-924b-3f14aeceeff8\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.674027 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-svc\") pod \"a7490f58-8a80-453b-924b-3f14aeceeff8\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.674633 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-nb\") pod \"a7490f58-8a80-453b-924b-3f14aeceeff8\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.674676 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-swift-storage-0\") pod \"a7490f58-8a80-453b-924b-3f14aeceeff8\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.674724 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwhvd\" (UniqueName: \"kubernetes.io/projected/a7490f58-8a80-453b-924b-3f14aeceeff8-kube-api-access-fwhvd\") pod \"a7490f58-8a80-453b-924b-3f14aeceeff8\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.674763 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-sb\") pod \"a7490f58-8a80-453b-924b-3f14aeceeff8\" (UID: \"a7490f58-8a80-453b-924b-3f14aeceeff8\") " Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.690244 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7490f58-8a80-453b-924b-3f14aeceeff8-kube-api-access-fwhvd" (OuterVolumeSpecName: "kube-api-access-fwhvd") pod "a7490f58-8a80-453b-924b-3f14aeceeff8" (UID: "a7490f58-8a80-453b-924b-3f14aeceeff8"). InnerVolumeSpecName "kube-api-access-fwhvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.691086 4715 scope.go:117] "RemoveContainer" containerID="031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927" Dec 10 09:56:30 crc kubenswrapper[4715]: E1210 09:56:30.691585 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927\": container with ID starting with 031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927 not found: ID does not exist" containerID="031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.691688 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927"} err="failed to get container status \"031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927\": rpc error: code = NotFound desc = could not find container \"031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927\": container with ID starting with 031853aed5acf155f78c15fb10c9048b603248bd43c088d802fb642b4c1e4927 not found: ID does not exist" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.691720 4715 scope.go:117] "RemoveContainer" containerID="5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837" Dec 10 09:56:30 crc kubenswrapper[4715]: E1210 09:56:30.692043 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837\": container with ID starting with 5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837 not found: ID does not exist" containerID="5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.692194 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837"} err="failed to get container status \"5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837\": rpc error: code = NotFound desc = could not find container \"5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837\": container with ID starting with 5360eed73b8a57f015ed8f4f4a15724982fca2970a07e463ba44f49b91c90837 not found: ID does not exist" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.726509 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a7490f58-8a80-453b-924b-3f14aeceeff8" (UID: "a7490f58-8a80-453b-924b-3f14aeceeff8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.731338 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7490f58-8a80-453b-924b-3f14aeceeff8" (UID: "a7490f58-8a80-453b-924b-3f14aeceeff8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.731472 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a7490f58-8a80-453b-924b-3f14aeceeff8" (UID: "a7490f58-8a80-453b-924b-3f14aeceeff8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.734705 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-config" (OuterVolumeSpecName: "config") pod "a7490f58-8a80-453b-924b-3f14aeceeff8" (UID: "a7490f58-8a80-453b-924b-3f14aeceeff8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.736126 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a7490f58-8a80-453b-924b-3f14aeceeff8" (UID: "a7490f58-8a80-453b-924b-3f14aeceeff8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.777001 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwhvd\" (UniqueName: \"kubernetes.io/projected/a7490f58-8a80-453b-924b-3f14aeceeff8-kube-api-access-fwhvd\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.777027 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.777036 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.777045 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.777053 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.777062 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7490f58-8a80-453b-924b-3f14aeceeff8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:30 crc kubenswrapper[4715]: I1210 09:56:30.780223 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-g755x"] Dec 10 09:56:31 crc kubenswrapper[4715]: I1210 09:56:31.000471 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fcqpx"] Dec 10 09:56:31 crc kubenswrapper[4715]: I1210 09:56:31.017817 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-fcqpx"] Dec 10 09:56:31 crc kubenswrapper[4715]: E1210 09:56:31.164300 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7490f58_8a80_453b_924b_3f14aeceeff8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod540bd567_e127_4db2_b6c5_0d447ce7b34c.slice/crio-conmon-39e4f2f2c82e00666d8ad4a4563247b5e7587ef5544bf7b824e2a0dc02d0a001.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7490f58_8a80_453b_924b_3f14aeceeff8.slice/crio-40c786517981b7605b57118882b6fef08690177aca913f5108808fb0871d09bc\": RecentStats: unable to find data in memory cache]" Dec 10 09:56:31 crc kubenswrapper[4715]: I1210 09:56:31.614730 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7490f58-8a80-453b-924b-3f14aeceeff8" path="/var/lib/kubelet/pods/a7490f58-8a80-453b-924b-3f14aeceeff8/volumes" Dec 10 09:56:31 crc kubenswrapper[4715]: I1210 09:56:31.637311 4715 generic.go:334] "Generic (PLEG): container finished" podID="540bd567-e127-4db2-b6c5-0d447ce7b34c" containerID="39e4f2f2c82e00666d8ad4a4563247b5e7587ef5544bf7b824e2a0dc02d0a001" exitCode=0 Dec 10 09:56:31 crc kubenswrapper[4715]: I1210 09:56:31.637354 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-g755x" event={"ID":"540bd567-e127-4db2-b6c5-0d447ce7b34c","Type":"ContainerDied","Data":"39e4f2f2c82e00666d8ad4a4563247b5e7587ef5544bf7b824e2a0dc02d0a001"} Dec 10 09:56:31 crc kubenswrapper[4715]: I1210 09:56:31.637393 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-g755x" event={"ID":"540bd567-e127-4db2-b6c5-0d447ce7b34c","Type":"ContainerStarted","Data":"80e27afcbf1f2ab2e790bfa0a77d6a0b4b79955b5f64cf8caed49293f10b9e49"} Dec 10 09:56:32 crc kubenswrapper[4715]: I1210 09:56:32.651598 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-g755x" event={"ID":"540bd567-e127-4db2-b6c5-0d447ce7b34c","Type":"ContainerStarted","Data":"1b5f78c4f4213aca86a68b92372dba2b1c96ac5f6ec6612bec0f942ff1787b43"} Dec 10 09:56:32 crc kubenswrapper[4715]: I1210 09:56:32.652054 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:32 crc kubenswrapper[4715]: I1210 09:56:32.671500 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-g755x" podStartSLOduration=3.6714812439999998 podStartE2EDuration="3.671481244s" podCreationTimestamp="2025-12-10 09:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:56:32.668347813 +0000 UTC m=+1355.411894084" watchObservedRunningTime="2025-12-10 09:56:32.671481244 +0000 UTC m=+1355.415027505" Dec 10 09:56:40 crc kubenswrapper[4715]: I1210 09:56:40.277057 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-g755x" Dec 10 09:56:40 crc kubenswrapper[4715]: I1210 09:56:40.342845 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lj6xg"] Dec 10 09:56:40 crc kubenswrapper[4715]: I1210 09:56:40.343120 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" podUID="ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" containerName="dnsmasq-dns" containerID="cri-o://5c956ce1b9205394740d68ef76851a5205f2a2d5a764ed3011ff9f1098ddafe6" gracePeriod=10 Dec 10 09:56:40 crc kubenswrapper[4715]: I1210 09:56:40.759829 4715 generic.go:334] "Generic (PLEG): container finished" podID="ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" containerID="5c956ce1b9205394740d68ef76851a5205f2a2d5a764ed3011ff9f1098ddafe6" exitCode=0 Dec 10 09:56:40 crc kubenswrapper[4715]: I1210 09:56:40.759938 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" event={"ID":"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41","Type":"ContainerDied","Data":"5c956ce1b9205394740d68ef76851a5205f2a2d5a764ed3011ff9f1098ddafe6"} Dec 10 09:56:40 crc kubenswrapper[4715]: I1210 09:56:40.869481 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.012387 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-svc\") pod \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.012458 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-config\") pod \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.012496 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5vjs\" (UniqueName: \"kubernetes.io/projected/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-kube-api-access-d5vjs\") pod \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.012538 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-openstack-edpm-ipam\") pod \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.012628 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-sb\") pod \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.012672 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-nb\") pod \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.012893 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-swift-storage-0\") pod \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\" (UID: \"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41\") " Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.028238 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-kube-api-access-d5vjs" (OuterVolumeSpecName: "kube-api-access-d5vjs") pod "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" (UID: "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41"). InnerVolumeSpecName "kube-api-access-d5vjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.078620 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" (UID: "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.086805 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" (UID: "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.089877 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-config" (OuterVolumeSpecName: "config") pod "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" (UID: "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.096887 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" (UID: "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.103709 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" (UID: "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.115566 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.115612 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-config\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.115625 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5vjs\" (UniqueName: \"kubernetes.io/projected/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-kube-api-access-d5vjs\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.115641 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.115653 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.115664 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.131298 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" (UID: "ff7c6489-b6c9-4c3b-8a96-c1ec22021e41"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.218283 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.770927 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" event={"ID":"ff7c6489-b6c9-4c3b-8a96-c1ec22021e41","Type":"ContainerDied","Data":"3d85f4d8c385804509745de68ffedb439fbd29781e10ebabe15a35cc0fbc8cac"} Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.770991 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-lj6xg" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.771404 4715 scope.go:117] "RemoveContainer" containerID="5c956ce1b9205394740d68ef76851a5205f2a2d5a764ed3011ff9f1098ddafe6" Dec 10 09:56:41 crc kubenswrapper[4715]: I1210 09:56:41.792726 4715 scope.go:117] "RemoveContainer" containerID="c4b06541705dc8c65df31fb1d289723fb64c1ed04324f08f37dac45fb5bbbcb9" Dec 10 09:56:42 crc kubenswrapper[4715]: I1210 09:56:42.170932 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lj6xg"] Dec 10 09:56:42 crc kubenswrapper[4715]: I1210 09:56:42.187675 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lj6xg"] Dec 10 09:56:43 crc kubenswrapper[4715]: I1210 09:56:43.618755 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" path="/var/lib/kubelet/pods/ff7c6489-b6c9-4c3b-8a96-c1ec22021e41/volumes" Dec 10 09:56:47 crc kubenswrapper[4715]: I1210 09:56:47.713997 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:56:47 crc kubenswrapper[4715]: I1210 09:56:47.714344 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:56:50 crc kubenswrapper[4715]: I1210 09:56:50.856993 4715 generic.go:334] "Generic (PLEG): container finished" podID="e4e8933c-d768-48cb-bac7-245abd8761ad" containerID="185ac203d8b588fc81ceeecce1f460e5ca65ffe6f0faa8d57781e308d4fe5c7c" exitCode=0 Dec 10 09:56:50 crc kubenswrapper[4715]: I1210 09:56:50.857400 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4e8933c-d768-48cb-bac7-245abd8761ad","Type":"ContainerDied","Data":"185ac203d8b588fc81ceeecce1f460e5ca65ffe6f0faa8d57781e308d4fe5c7c"} Dec 10 09:56:51 crc kubenswrapper[4715]: I1210 09:56:51.872317 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4e8933c-d768-48cb-bac7-245abd8761ad","Type":"ContainerStarted","Data":"509f7f3e91c19b9e175509e5a0c3f4b7b43f17ab36148b067b364b14f3a18cf7"} Dec 10 09:56:51 crc kubenswrapper[4715]: I1210 09:56:51.874576 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 10 09:56:51 crc kubenswrapper[4715]: I1210 09:56:51.877335 4715 generic.go:334] "Generic (PLEG): container finished" podID="bf6962bf-0a77-4285-b58f-6213688e74b7" containerID="c4fabcfa33893a06e4684442f7b7a4379b662d870d5b75e0da31be85d2b2482e" exitCode=0 Dec 10 09:56:51 crc kubenswrapper[4715]: I1210 09:56:51.877387 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf6962bf-0a77-4285-b58f-6213688e74b7","Type":"ContainerDied","Data":"c4fabcfa33893a06e4684442f7b7a4379b662d870d5b75e0da31be85d2b2482e"} Dec 10 09:56:51 crc kubenswrapper[4715]: I1210 09:56:51.914423 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.914401948 podStartE2EDuration="36.914401948s" podCreationTimestamp="2025-12-10 09:56:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:56:51.905462868 +0000 UTC m=+1374.649009119" watchObservedRunningTime="2025-12-10 09:56:51.914401948 +0000 UTC m=+1374.657948219" Dec 10 09:56:52 crc kubenswrapper[4715]: I1210 09:56:52.887982 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf6962bf-0a77-4285-b58f-6213688e74b7","Type":"ContainerStarted","Data":"cae6231daabdc96f2c95000dafb1f2d8fe69bdcae5a9bca18cc1eef3cc41c158"} Dec 10 09:56:52 crc kubenswrapper[4715]: I1210 09:56:52.913072 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.913048978 podStartE2EDuration="36.913048978s" podCreationTimestamp="2025-12-10 09:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 09:56:52.909745992 +0000 UTC m=+1375.653292253" watchObservedRunningTime="2025-12-10 09:56:52.913048978 +0000 UTC m=+1375.656595229" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.454253 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl"] Dec 10 09:56:53 crc kubenswrapper[4715]: E1210 09:56:53.455039 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7490f58-8a80-453b-924b-3f14aeceeff8" containerName="init" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.455055 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7490f58-8a80-453b-924b-3f14aeceeff8" containerName="init" Dec 10 09:56:53 crc kubenswrapper[4715]: E1210 09:56:53.455091 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" containerName="dnsmasq-dns" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.455097 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" containerName="dnsmasq-dns" Dec 10 09:56:53 crc kubenswrapper[4715]: E1210 09:56:53.455113 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" containerName="init" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.455119 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" containerName="init" Dec 10 09:56:53 crc kubenswrapper[4715]: E1210 09:56:53.455130 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7490f58-8a80-453b-924b-3f14aeceeff8" containerName="dnsmasq-dns" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.455137 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7490f58-8a80-453b-924b-3f14aeceeff8" containerName="dnsmasq-dns" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.455338 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7490f58-8a80-453b-924b-3f14aeceeff8" containerName="dnsmasq-dns" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.455353 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7c6489-b6c9-4c3b-8a96-c1ec22021e41" containerName="dnsmasq-dns" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.456156 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.458619 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.458787 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.459186 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.459769 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.473568 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl"] Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.591314 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb4z2\" (UniqueName: \"kubernetes.io/projected/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-kube-api-access-sb4z2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.591369 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.591585 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.591655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.694363 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.694457 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.694605 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb4z2\" (UniqueName: \"kubernetes.io/projected/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-kube-api-access-sb4z2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.694696 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.699592 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.700377 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.703593 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.713847 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb4z2\" (UniqueName: \"kubernetes.io/projected/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-kube-api-access-sb4z2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:53 crc kubenswrapper[4715]: I1210 09:56:53.782095 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:56:54 crc kubenswrapper[4715]: I1210 09:56:54.537506 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl"] Dec 10 09:56:54 crc kubenswrapper[4715]: I1210 09:56:54.914935 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" event={"ID":"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd","Type":"ContainerStarted","Data":"23b1dd2cf07813944f9af858df59f3b82c3291bd711b376977273171080bcd3f"} Dec 10 09:56:57 crc kubenswrapper[4715]: I1210 09:56:57.087846 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:57:05 crc kubenswrapper[4715]: I1210 09:57:05.039331 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 09:57:05 crc kubenswrapper[4715]: I1210 09:57:05.994204 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 10 09:57:06 crc kubenswrapper[4715]: I1210 09:57:06.053900 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" event={"ID":"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd","Type":"ContainerStarted","Data":"3f6ec31a5cc20d5728995637e1c3a0a8a44a63a4969cd464d0b7812e736171d0"} Dec 10 09:57:07 crc kubenswrapper[4715]: I1210 09:57:07.091103 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 10 09:57:07 crc kubenswrapper[4715]: I1210 09:57:07.125041 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" podStartSLOduration=3.635981475 podStartE2EDuration="14.125015895s" podCreationTimestamp="2025-12-10 09:56:53 +0000 UTC" firstStartedPulling="2025-12-10 09:56:54.544606509 +0000 UTC m=+1377.288152750" lastFinishedPulling="2025-12-10 09:57:05.033640919 +0000 UTC m=+1387.777187170" observedRunningTime="2025-12-10 09:57:06.082570722 +0000 UTC m=+1388.826116983" watchObservedRunningTime="2025-12-10 09:57:07.125015895 +0000 UTC m=+1389.868562146" Dec 10 09:57:17 crc kubenswrapper[4715]: I1210 09:57:17.150345 4715 generic.go:334] "Generic (PLEG): container finished" podID="17c08b14-a1de-4e79-b6d6-a8543bb0a4fd" containerID="3f6ec31a5cc20d5728995637e1c3a0a8a44a63a4969cd464d0b7812e736171d0" exitCode=0 Dec 10 09:57:17 crc kubenswrapper[4715]: I1210 09:57:17.150417 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" event={"ID":"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd","Type":"ContainerDied","Data":"3f6ec31a5cc20d5728995637e1c3a0a8a44a63a4969cd464d0b7812e736171d0"} Dec 10 09:57:17 crc kubenswrapper[4715]: I1210 09:57:17.714353 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:57:17 crc kubenswrapper[4715]: I1210 09:57:17.714420 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.546261 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.565189 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-inventory\") pod \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.565548 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-ssh-key\") pod \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.565702 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-repo-setup-combined-ca-bundle\") pod \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.565743 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb4z2\" (UniqueName: \"kubernetes.io/projected/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-kube-api-access-sb4z2\") pod \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\" (UID: \"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd\") " Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.571991 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-kube-api-access-sb4z2" (OuterVolumeSpecName: "kube-api-access-sb4z2") pod "17c08b14-a1de-4e79-b6d6-a8543bb0a4fd" (UID: "17c08b14-a1de-4e79-b6d6-a8543bb0a4fd"). InnerVolumeSpecName "kube-api-access-sb4z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.572822 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "17c08b14-a1de-4e79-b6d6-a8543bb0a4fd" (UID: "17c08b14-a1de-4e79-b6d6-a8543bb0a4fd"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.598301 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "17c08b14-a1de-4e79-b6d6-a8543bb0a4fd" (UID: "17c08b14-a1de-4e79-b6d6-a8543bb0a4fd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.604065 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-inventory" (OuterVolumeSpecName: "inventory") pod "17c08b14-a1de-4e79-b6d6-a8543bb0a4fd" (UID: "17c08b14-a1de-4e79-b6d6-a8543bb0a4fd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.668381 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb4z2\" (UniqueName: \"kubernetes.io/projected/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-kube-api-access-sb4z2\") on node \"crc\" DevicePath \"\"" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.668415 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.668424 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:18.668433 4715 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c08b14-a1de-4e79-b6d6-a8543bb0a4fd-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.169087 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" event={"ID":"17c08b14-a1de-4e79-b6d6-a8543bb0a4fd","Type":"ContainerDied","Data":"23b1dd2cf07813944f9af858df59f3b82c3291bd711b376977273171080bcd3f"} Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.169127 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23b1dd2cf07813944f9af858df59f3b82c3291bd711b376977273171080bcd3f" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.169137 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.251962 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7"] Dec 10 09:57:19 crc kubenswrapper[4715]: E1210 09:57:19.252509 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17c08b14-a1de-4e79-b6d6-a8543bb0a4fd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.252535 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="17c08b14-a1de-4e79-b6d6-a8543bb0a4fd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.252759 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="17c08b14-a1de-4e79-b6d6-a8543bb0a4fd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.253668 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.262979 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.263192 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.263198 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.263362 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.271423 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7"] Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.281577 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rcr7\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.281691 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rcr7\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.281762 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-824cb\" (UniqueName: \"kubernetes.io/projected/bd4fa444-6ad3-4129-b62d-c98ce1b69058-kube-api-access-824cb\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rcr7\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.383995 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-824cb\" (UniqueName: \"kubernetes.io/projected/bd4fa444-6ad3-4129-b62d-c98ce1b69058-kube-api-access-824cb\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rcr7\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.384138 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rcr7\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.384240 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rcr7\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.388636 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rcr7\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.390822 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rcr7\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.400645 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-824cb\" (UniqueName: \"kubernetes.io/projected/bd4fa444-6ad3-4129-b62d-c98ce1b69058-kube-api-access-824cb\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5rcr7\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:19 crc kubenswrapper[4715]: I1210 09:57:19.583742 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:20 crc kubenswrapper[4715]: I1210 09:57:20.132813 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7"] Dec 10 09:57:20 crc kubenswrapper[4715]: I1210 09:57:20.179867 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" event={"ID":"bd4fa444-6ad3-4129-b62d-c98ce1b69058","Type":"ContainerStarted","Data":"5d840572c1855fe461f84ba33d665b336d13cd69d4b261782c1bd1d0b8f92f7b"} Dec 10 09:57:21 crc kubenswrapper[4715]: I1210 09:57:21.192433 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" event={"ID":"bd4fa444-6ad3-4129-b62d-c98ce1b69058","Type":"ContainerStarted","Data":"9afbe6393d7db5f7b310048b2fe6e9d27d874c21a207f8e91e30b5a47f4a3bd0"} Dec 10 09:57:21 crc kubenswrapper[4715]: I1210 09:57:21.213724 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" podStartSLOduration=1.809938211 podStartE2EDuration="2.213706336s" podCreationTimestamp="2025-12-10 09:57:19 +0000 UTC" firstStartedPulling="2025-12-10 09:57:20.134878644 +0000 UTC m=+1402.878424895" lastFinishedPulling="2025-12-10 09:57:20.538646769 +0000 UTC m=+1403.282193020" observedRunningTime="2025-12-10 09:57:21.205799806 +0000 UTC m=+1403.949346057" watchObservedRunningTime="2025-12-10 09:57:21.213706336 +0000 UTC m=+1403.957252577" Dec 10 09:57:24 crc kubenswrapper[4715]: I1210 09:57:24.222428 4715 generic.go:334] "Generic (PLEG): container finished" podID="bd4fa444-6ad3-4129-b62d-c98ce1b69058" containerID="9afbe6393d7db5f7b310048b2fe6e9d27d874c21a207f8e91e30b5a47f4a3bd0" exitCode=0 Dec 10 09:57:24 crc kubenswrapper[4715]: I1210 09:57:24.222498 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" event={"ID":"bd4fa444-6ad3-4129-b62d-c98ce1b69058","Type":"ContainerDied","Data":"9afbe6393d7db5f7b310048b2fe6e9d27d874c21a207f8e91e30b5a47f4a3bd0"} Dec 10 09:57:25 crc kubenswrapper[4715]: I1210 09:57:25.662649 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:25 crc kubenswrapper[4715]: I1210 09:57:25.811774 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-inventory\") pod \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " Dec 10 09:57:25 crc kubenswrapper[4715]: I1210 09:57:25.811934 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-ssh-key\") pod \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " Dec 10 09:57:25 crc kubenswrapper[4715]: I1210 09:57:25.812184 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-824cb\" (UniqueName: \"kubernetes.io/projected/bd4fa444-6ad3-4129-b62d-c98ce1b69058-kube-api-access-824cb\") pod \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\" (UID: \"bd4fa444-6ad3-4129-b62d-c98ce1b69058\") " Dec 10 09:57:25 crc kubenswrapper[4715]: I1210 09:57:25.837842 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd4fa444-6ad3-4129-b62d-c98ce1b69058-kube-api-access-824cb" (OuterVolumeSpecName: "kube-api-access-824cb") pod "bd4fa444-6ad3-4129-b62d-c98ce1b69058" (UID: "bd4fa444-6ad3-4129-b62d-c98ce1b69058"). InnerVolumeSpecName "kube-api-access-824cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 09:57:25 crc kubenswrapper[4715]: I1210 09:57:25.848212 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-inventory" (OuterVolumeSpecName: "inventory") pod "bd4fa444-6ad3-4129-b62d-c98ce1b69058" (UID: "bd4fa444-6ad3-4129-b62d-c98ce1b69058"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:57:25 crc kubenswrapper[4715]: I1210 09:57:25.864496 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bd4fa444-6ad3-4129-b62d-c98ce1b69058" (UID: "bd4fa444-6ad3-4129-b62d-c98ce1b69058"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 09:57:25 crc kubenswrapper[4715]: I1210 09:57:25.914470 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 09:57:25 crc kubenswrapper[4715]: I1210 09:57:25.914495 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd4fa444-6ad3-4129-b62d-c98ce1b69058-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 09:57:25 crc kubenswrapper[4715]: I1210 09:57:25.914507 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-824cb\" (UniqueName: \"kubernetes.io/projected/bd4fa444-6ad3-4129-b62d-c98ce1b69058-kube-api-access-824cb\") on node \"crc\" DevicePath \"\"" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.244770 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" event={"ID":"bd4fa444-6ad3-4129-b62d-c98ce1b69058","Type":"ContainerDied","Data":"5d840572c1855fe461f84ba33d665b336d13cd69d4b261782c1bd1d0b8f92f7b"} Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.244845 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d840572c1855fe461f84ba33d665b336d13cd69d4b261782c1bd1d0b8f92f7b" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.244870 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5rcr7" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.316660 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw"] Dec 10 09:57:26 crc kubenswrapper[4715]: E1210 09:57:26.317263 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd4fa444-6ad3-4129-b62d-c98ce1b69058" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.317294 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd4fa444-6ad3-4129-b62d-c98ce1b69058" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.317588 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd4fa444-6ad3-4129-b62d-c98ce1b69058" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.318475 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.320747 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.321690 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.322602 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.323013 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.326601 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw"] Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.424495 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.424585 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrfzv\" (UniqueName: \"kubernetes.io/projected/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-kube-api-access-jrfzv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.424626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.424719 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.527253 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.527322 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrfzv\" (UniqueName: \"kubernetes.io/projected/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-kube-api-access-jrfzv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.527354 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.527479 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.532314 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.532504 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.533744 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.550444 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrfzv\" (UniqueName: \"kubernetes.io/projected/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-kube-api-access-jrfzv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:26 crc kubenswrapper[4715]: I1210 09:57:26.639084 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 09:57:27 crc kubenswrapper[4715]: I1210 09:57:27.229549 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw"] Dec 10 09:57:27 crc kubenswrapper[4715]: W1210 09:57:27.232813 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bcd9508_4ca1_4bff_9e27_9c3083ee0ae2.slice/crio-9253df8af2b519313d9b816bb334933dcd87ba8d8dc1ea4f8480a96184aac6a8 WatchSource:0}: Error finding container 9253df8af2b519313d9b816bb334933dcd87ba8d8dc1ea4f8480a96184aac6a8: Status 404 returned error can't find the container with id 9253df8af2b519313d9b816bb334933dcd87ba8d8dc1ea4f8480a96184aac6a8 Dec 10 09:57:27 crc kubenswrapper[4715]: I1210 09:57:27.255760 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" event={"ID":"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2","Type":"ContainerStarted","Data":"9253df8af2b519313d9b816bb334933dcd87ba8d8dc1ea4f8480a96184aac6a8"} Dec 10 09:57:28 crc kubenswrapper[4715]: I1210 09:57:28.267263 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" event={"ID":"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2","Type":"ContainerStarted","Data":"65161dd09051dccc1fb6db068c1daabf301164ad7d38c22f385b83c813b07257"} Dec 10 09:57:28 crc kubenswrapper[4715]: I1210 09:57:28.286716 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" podStartSLOduration=1.849658862 podStartE2EDuration="2.286694575s" podCreationTimestamp="2025-12-10 09:57:26 +0000 UTC" firstStartedPulling="2025-12-10 09:57:27.235801856 +0000 UTC m=+1409.979348117" lastFinishedPulling="2025-12-10 09:57:27.672837579 +0000 UTC m=+1410.416383830" observedRunningTime="2025-12-10 09:57:28.283597405 +0000 UTC m=+1411.027143666" watchObservedRunningTime="2025-12-10 09:57:28.286694575 +0000 UTC m=+1411.030240816" Dec 10 09:57:47 crc kubenswrapper[4715]: I1210 09:57:47.713728 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 09:57:47 crc kubenswrapper[4715]: I1210 09:57:47.714302 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 09:57:47 crc kubenswrapper[4715]: I1210 09:57:47.714347 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 09:57:47 crc kubenswrapper[4715]: I1210 09:57:47.715453 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"43f0315f5b714d86cf23ed6f6c468531cbdb76f130137362d46c5a9fc1556442"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 09:57:47 crc kubenswrapper[4715]: I1210 09:57:47.715529 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://43f0315f5b714d86cf23ed6f6c468531cbdb76f130137362d46c5a9fc1556442" gracePeriod=600 Dec 10 09:57:48 crc kubenswrapper[4715]: I1210 09:57:48.734052 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="43f0315f5b714d86cf23ed6f6c468531cbdb76f130137362d46c5a9fc1556442" exitCode=0 Dec 10 09:57:48 crc kubenswrapper[4715]: I1210 09:57:48.734117 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"43f0315f5b714d86cf23ed6f6c468531cbdb76f130137362d46c5a9fc1556442"} Dec 10 09:57:48 crc kubenswrapper[4715]: I1210 09:57:48.735046 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7"} Dec 10 09:57:48 crc kubenswrapper[4715]: I1210 09:57:48.735080 4715 scope.go:117] "RemoveContainer" containerID="13f540c3510be87c668f6f3667fb9d329e2ff72443e0f54c1a4f373c716a3669" Dec 10 09:58:21 crc kubenswrapper[4715]: I1210 09:58:21.067762 4715 scope.go:117] "RemoveContainer" containerID="e3c748c38073b710935f8bc8f46b6ae6913e36215588881f9fdd6733d13a7b41" Dec 10 09:58:21 crc kubenswrapper[4715]: I1210 09:58:21.108798 4715 scope.go:117] "RemoveContainer" containerID="7fe183b9869687758f0327fc261359e2806380711d2ad774c34e9dcf7fc8424a" Dec 10 09:58:21 crc kubenswrapper[4715]: I1210 09:58:21.156021 4715 scope.go:117] "RemoveContainer" containerID="34e5c1b0be35fb91bfd6fc9c791e3e819c821cd7fd272cab546dde7efdec1e10" Dec 10 09:58:21 crc kubenswrapper[4715]: I1210 09:58:21.193595 4715 scope.go:117] "RemoveContainer" containerID="4bc5cf13ef13df14c971a186ec8641765e3aa9c68ef0332d60027c065e9b6aa9" Dec 10 09:58:21 crc kubenswrapper[4715]: I1210 09:58:21.248941 4715 scope.go:117] "RemoveContainer" containerID="a6a3751858a858b59e78c6e794614c1bb47d642d5cd6c2c664dee03ecd418eb9" Dec 10 09:59:21 crc kubenswrapper[4715]: I1210 09:59:21.354872 4715 scope.go:117] "RemoveContainer" containerID="51e47c5a22fe27694a01c0e2b0c68edc82cd7c5c052747e4d8a03a8164dbf545" Dec 10 09:59:21 crc kubenswrapper[4715]: I1210 09:59:21.391718 4715 scope.go:117] "RemoveContainer" containerID="77d73f827aff7bd2ee4d514c4e9756d7255049a02f1d032a89c5b6b1b046e946" Dec 10 09:59:21 crc kubenswrapper[4715]: I1210 09:59:21.413977 4715 scope.go:117] "RemoveContainer" containerID="c3a435ba5c7d59aec6834021bd7ea496b620189b0bcf79a4db85af3b4e6c81de" Dec 10 09:59:21 crc kubenswrapper[4715]: I1210 09:59:21.468335 4715 scope.go:117] "RemoveContainer" containerID="8ff58dd4c00af545e2be2e5ebc3e656b8e492796b8d42a11c5a8b4dbbb62373b" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.476759 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-46q9k"] Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.479872 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.494328 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-46q9k"] Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.638042 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-catalog-content\") pod \"community-operators-46q9k\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.638190 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx76g\" (UniqueName: \"kubernetes.io/projected/7c9fe615-64aa-490f-b79d-63a032c0b219-kube-api-access-rx76g\") pod \"community-operators-46q9k\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.638211 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-utilities\") pod \"community-operators-46q9k\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.744042 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx76g\" (UniqueName: \"kubernetes.io/projected/7c9fe615-64aa-490f-b79d-63a032c0b219-kube-api-access-rx76g\") pod \"community-operators-46q9k\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.744093 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-utilities\") pod \"community-operators-46q9k\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.744231 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-catalog-content\") pod \"community-operators-46q9k\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.746907 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-catalog-content\") pod \"community-operators-46q9k\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.746999 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-utilities\") pod \"community-operators-46q9k\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.782298 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx76g\" (UniqueName: \"kubernetes.io/projected/7c9fe615-64aa-490f-b79d-63a032c0b219-kube-api-access-rx76g\") pod \"community-operators-46q9k\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:47 crc kubenswrapper[4715]: I1210 09:59:47.806418 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:48 crc kubenswrapper[4715]: I1210 09:59:48.426884 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-46q9k"] Dec 10 09:59:49 crc kubenswrapper[4715]: I1210 09:59:49.170991 4715 generic.go:334] "Generic (PLEG): container finished" podID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerID="7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20" exitCode=0 Dec 10 09:59:49 crc kubenswrapper[4715]: I1210 09:59:49.171140 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-46q9k" event={"ID":"7c9fe615-64aa-490f-b79d-63a032c0b219","Type":"ContainerDied","Data":"7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20"} Dec 10 09:59:49 crc kubenswrapper[4715]: I1210 09:59:49.171341 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-46q9k" event={"ID":"7c9fe615-64aa-490f-b79d-63a032c0b219","Type":"ContainerStarted","Data":"dc4edebff8d3e85bc63223a3b6fddf9e2c0ab9db0ca35c0681e7e307079eafe8"} Dec 10 09:59:49 crc kubenswrapper[4715]: I1210 09:59:49.173332 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 09:59:51 crc kubenswrapper[4715]: I1210 09:59:51.192238 4715 generic.go:334] "Generic (PLEG): container finished" podID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerID="72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de" exitCode=0 Dec 10 09:59:51 crc kubenswrapper[4715]: I1210 09:59:51.192348 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-46q9k" event={"ID":"7c9fe615-64aa-490f-b79d-63a032c0b219","Type":"ContainerDied","Data":"72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de"} Dec 10 09:59:52 crc kubenswrapper[4715]: I1210 09:59:52.203666 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-46q9k" event={"ID":"7c9fe615-64aa-490f-b79d-63a032c0b219","Type":"ContainerStarted","Data":"2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e"} Dec 10 09:59:52 crc kubenswrapper[4715]: I1210 09:59:52.233805 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-46q9k" podStartSLOduration=2.693124664 podStartE2EDuration="5.233780788s" podCreationTimestamp="2025-12-10 09:59:47 +0000 UTC" firstStartedPulling="2025-12-10 09:59:49.173104561 +0000 UTC m=+1551.916650812" lastFinishedPulling="2025-12-10 09:59:51.713760685 +0000 UTC m=+1554.457306936" observedRunningTime="2025-12-10 09:59:52.225562062 +0000 UTC m=+1554.969108323" watchObservedRunningTime="2025-12-10 09:59:52.233780788 +0000 UTC m=+1554.977327049" Dec 10 09:59:57 crc kubenswrapper[4715]: I1210 09:59:57.807963 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:57 crc kubenswrapper[4715]: I1210 09:59:57.808479 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:57 crc kubenswrapper[4715]: I1210 09:59:57.883650 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:58 crc kubenswrapper[4715]: I1210 09:59:58.322063 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-46q9k" Dec 10 09:59:58 crc kubenswrapper[4715]: I1210 09:59:58.385516 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-46q9k"] Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.152520 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4"] Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.154987 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.158778 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.161009 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.163971 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4"] Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.275436 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-46q9k" podUID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerName="registry-server" containerID="cri-o://2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e" gracePeriod=2 Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.316563 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb60d0d6-63d3-4453-9ae4-365ff52a48de-secret-volume\") pod \"collect-profiles-29422680-6hjk4\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.316638 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxd86\" (UniqueName: \"kubernetes.io/projected/eb60d0d6-63d3-4453-9ae4-365ff52a48de-kube-api-access-mxd86\") pod \"collect-profiles-29422680-6hjk4\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.316850 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb60d0d6-63d3-4453-9ae4-365ff52a48de-config-volume\") pod \"collect-profiles-29422680-6hjk4\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.418337 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxd86\" (UniqueName: \"kubernetes.io/projected/eb60d0d6-63d3-4453-9ae4-365ff52a48de-kube-api-access-mxd86\") pod \"collect-profiles-29422680-6hjk4\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.418470 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb60d0d6-63d3-4453-9ae4-365ff52a48de-config-volume\") pod \"collect-profiles-29422680-6hjk4\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.418545 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb60d0d6-63d3-4453-9ae4-365ff52a48de-secret-volume\") pod \"collect-profiles-29422680-6hjk4\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.419652 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb60d0d6-63d3-4453-9ae4-365ff52a48de-config-volume\") pod \"collect-profiles-29422680-6hjk4\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.424265 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb60d0d6-63d3-4453-9ae4-365ff52a48de-secret-volume\") pod \"collect-profiles-29422680-6hjk4\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.440059 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxd86\" (UniqueName: \"kubernetes.io/projected/eb60d0d6-63d3-4453-9ae4-365ff52a48de-kube-api-access-mxd86\") pod \"collect-profiles-29422680-6hjk4\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.490675 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.811612 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-46q9k" Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.986558 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-catalog-content\") pod \"7c9fe615-64aa-490f-b79d-63a032c0b219\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.986716 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-utilities\") pod \"7c9fe615-64aa-490f-b79d-63a032c0b219\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.986755 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rx76g\" (UniqueName: \"kubernetes.io/projected/7c9fe615-64aa-490f-b79d-63a032c0b219-kube-api-access-rx76g\") pod \"7c9fe615-64aa-490f-b79d-63a032c0b219\" (UID: \"7c9fe615-64aa-490f-b79d-63a032c0b219\") " Dec 10 10:00:00 crc kubenswrapper[4715]: I1210 10:00:00.989671 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-utilities" (OuterVolumeSpecName: "utilities") pod "7c9fe615-64aa-490f-b79d-63a032c0b219" (UID: "7c9fe615-64aa-490f-b79d-63a032c0b219"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.015858 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c9fe615-64aa-490f-b79d-63a032c0b219-kube-api-access-rx76g" (OuterVolumeSpecName: "kube-api-access-rx76g") pod "7c9fe615-64aa-490f-b79d-63a032c0b219" (UID: "7c9fe615-64aa-490f-b79d-63a032c0b219"). InnerVolumeSpecName "kube-api-access-rx76g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.054521 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4"] Dec 10 10:00:01 crc kubenswrapper[4715]: W1210 10:00:01.056092 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb60d0d6_63d3_4453_9ae4_365ff52a48de.slice/crio-f0fb662cc5c5a84e0ad7a69cc5e48fe0974a3b9c05e21ed9f62f3183e3630445 WatchSource:0}: Error finding container f0fb662cc5c5a84e0ad7a69cc5e48fe0974a3b9c05e21ed9f62f3183e3630445: Status 404 returned error can't find the container with id f0fb662cc5c5a84e0ad7a69cc5e48fe0974a3b9c05e21ed9f62f3183e3630445 Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.089664 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.089707 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rx76g\" (UniqueName: \"kubernetes.io/projected/7c9fe615-64aa-490f-b79d-63a032c0b219-kube-api-access-rx76g\") on node \"crc\" DevicePath \"\"" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.288941 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" event={"ID":"eb60d0d6-63d3-4453-9ae4-365ff52a48de","Type":"ContainerStarted","Data":"f0fb662cc5c5a84e0ad7a69cc5e48fe0974a3b9c05e21ed9f62f3183e3630445"} Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.291602 4715 generic.go:334] "Generic (PLEG): container finished" podID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerID="2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e" exitCode=0 Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.291635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-46q9k" event={"ID":"7c9fe615-64aa-490f-b79d-63a032c0b219","Type":"ContainerDied","Data":"2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e"} Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.291653 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-46q9k" event={"ID":"7c9fe615-64aa-490f-b79d-63a032c0b219","Type":"ContainerDied","Data":"dc4edebff8d3e85bc63223a3b6fddf9e2c0ab9db0ca35c0681e7e307079eafe8"} Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.291672 4715 scope.go:117] "RemoveContainer" containerID="2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.291864 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-46q9k" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.316649 4715 scope.go:117] "RemoveContainer" containerID="72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.345404 4715 scope.go:117] "RemoveContainer" containerID="7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.364766 4715 scope.go:117] "RemoveContainer" containerID="2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e" Dec 10 10:00:01 crc kubenswrapper[4715]: E1210 10:00:01.365365 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e\": container with ID starting with 2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e not found: ID does not exist" containerID="2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.365400 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e"} err="failed to get container status \"2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e\": rpc error: code = NotFound desc = could not find container \"2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e\": container with ID starting with 2a024ac9051c5715756f1954fb961bf8e7da22dd1d5e3033fd358420644bdc7e not found: ID does not exist" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.365424 4715 scope.go:117] "RemoveContainer" containerID="72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de" Dec 10 10:00:01 crc kubenswrapper[4715]: E1210 10:00:01.365693 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de\": container with ID starting with 72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de not found: ID does not exist" containerID="72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.365721 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de"} err="failed to get container status \"72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de\": rpc error: code = NotFound desc = could not find container \"72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de\": container with ID starting with 72db127445dbf051719d84c97fe8b4607e9147ac7183464efd433416d2df84de not found: ID does not exist" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.365740 4715 scope.go:117] "RemoveContainer" containerID="7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20" Dec 10 10:00:01 crc kubenswrapper[4715]: E1210 10:00:01.366154 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20\": container with ID starting with 7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20 not found: ID does not exist" containerID="7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.366186 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20"} err="failed to get container status \"7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20\": rpc error: code = NotFound desc = could not find container \"7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20\": container with ID starting with 7e3e77ca9ca93f0560d65fb3c70b63cbf2cf32d6dfce042faa718e61e7f5ba20 not found: ID does not exist" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.478842 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c9fe615-64aa-490f-b79d-63a032c0b219" (UID: "7c9fe615-64aa-490f-b79d-63a032c0b219"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.497401 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c9fe615-64aa-490f-b79d-63a032c0b219-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.628000 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-46q9k"] Dec 10 10:00:01 crc kubenswrapper[4715]: I1210 10:00:01.643581 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-46q9k"] Dec 10 10:00:02 crc kubenswrapper[4715]: I1210 10:00:02.305960 4715 generic.go:334] "Generic (PLEG): container finished" podID="eb60d0d6-63d3-4453-9ae4-365ff52a48de" containerID="fcdb2b908a243bc0856c2c69201d5ff42d91c4c923a59777d9c1326f8f4a18d9" exitCode=0 Dec 10 10:00:02 crc kubenswrapper[4715]: I1210 10:00:02.306024 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" event={"ID":"eb60d0d6-63d3-4453-9ae4-365ff52a48de","Type":"ContainerDied","Data":"fcdb2b908a243bc0856c2c69201d5ff42d91c4c923a59777d9c1326f8f4a18d9"} Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.618554 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c9fe615-64aa-490f-b79d-63a032c0b219" path="/var/lib/kubelet/pods/7c9fe615-64aa-490f-b79d-63a032c0b219/volumes" Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.661950 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.842150 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxd86\" (UniqueName: \"kubernetes.io/projected/eb60d0d6-63d3-4453-9ae4-365ff52a48de-kube-api-access-mxd86\") pod \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.842198 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb60d0d6-63d3-4453-9ae4-365ff52a48de-config-volume\") pod \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.842245 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb60d0d6-63d3-4453-9ae4-365ff52a48de-secret-volume\") pod \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\" (UID: \"eb60d0d6-63d3-4453-9ae4-365ff52a48de\") " Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.842878 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb60d0d6-63d3-4453-9ae4-365ff52a48de-config-volume" (OuterVolumeSpecName: "config-volume") pod "eb60d0d6-63d3-4453-9ae4-365ff52a48de" (UID: "eb60d0d6-63d3-4453-9ae4-365ff52a48de"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.847840 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb60d0d6-63d3-4453-9ae4-365ff52a48de-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "eb60d0d6-63d3-4453-9ae4-365ff52a48de" (UID: "eb60d0d6-63d3-4453-9ae4-365ff52a48de"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.848074 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb60d0d6-63d3-4453-9ae4-365ff52a48de-kube-api-access-mxd86" (OuterVolumeSpecName: "kube-api-access-mxd86") pod "eb60d0d6-63d3-4453-9ae4-365ff52a48de" (UID: "eb60d0d6-63d3-4453-9ae4-365ff52a48de"). InnerVolumeSpecName "kube-api-access-mxd86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.944258 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxd86\" (UniqueName: \"kubernetes.io/projected/eb60d0d6-63d3-4453-9ae4-365ff52a48de-kube-api-access-mxd86\") on node \"crc\" DevicePath \"\"" Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.944293 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb60d0d6-63d3-4453-9ae4-365ff52a48de-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 10:00:03 crc kubenswrapper[4715]: I1210 10:00:03.944303 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb60d0d6-63d3-4453-9ae4-365ff52a48de-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 10:00:04 crc kubenswrapper[4715]: I1210 10:00:04.326897 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" event={"ID":"eb60d0d6-63d3-4453-9ae4-365ff52a48de","Type":"ContainerDied","Data":"f0fb662cc5c5a84e0ad7a69cc5e48fe0974a3b9c05e21ed9f62f3183e3630445"} Dec 10 10:00:04 crc kubenswrapper[4715]: I1210 10:00:04.326960 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0fb662cc5c5a84e0ad7a69cc5e48fe0974a3b9c05e21ed9f62f3183e3630445" Dec 10 10:00:04 crc kubenswrapper[4715]: I1210 10:00:04.326997 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4" Dec 10 10:00:17 crc kubenswrapper[4715]: I1210 10:00:17.713957 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:00:17 crc kubenswrapper[4715]: I1210 10:00:17.714542 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:00:21 crc kubenswrapper[4715]: I1210 10:00:21.579567 4715 scope.go:117] "RemoveContainer" containerID="d6302cf81087d77c068ea48b3cfd604814cda0c8abe9e8d0fe4a3c5a19a09b0a" Dec 10 10:00:21 crc kubenswrapper[4715]: I1210 10:00:21.604180 4715 scope.go:117] "RemoveContainer" containerID="093fc62679b5d45882383a16e1918725d935a37e09282cdbfecb2af62c10e7fd" Dec 10 10:00:21 crc kubenswrapper[4715]: I1210 10:00:21.624347 4715 scope.go:117] "RemoveContainer" containerID="fba9262b625db5aaf4e7f990bc5320905f4e4f9343dc14c54ea7f9c1f1f670ae" Dec 10 10:00:21 crc kubenswrapper[4715]: I1210 10:00:21.651043 4715 scope.go:117] "RemoveContainer" containerID="83d45601b0b9a2b50f6f032644d5a95641a14f5c9103d638a58d4edb179674eb" Dec 10 10:00:31 crc kubenswrapper[4715]: I1210 10:00:31.576246 4715 generic.go:334] "Generic (PLEG): container finished" podID="7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2" containerID="65161dd09051dccc1fb6db068c1daabf301164ad7d38c22f385b83c813b07257" exitCode=0 Dec 10 10:00:31 crc kubenswrapper[4715]: I1210 10:00:31.576321 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" event={"ID":"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2","Type":"ContainerDied","Data":"65161dd09051dccc1fb6db068c1daabf301164ad7d38c22f385b83c813b07257"} Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.030219 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.176803 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-inventory\") pod \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.177647 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-bootstrap-combined-ca-bundle\") pod \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.177729 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrfzv\" (UniqueName: \"kubernetes.io/projected/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-kube-api-access-jrfzv\") pod \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.177746 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-ssh-key\") pod \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\" (UID: \"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2\") " Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.184065 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2" (UID: "7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.184326 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-kube-api-access-jrfzv" (OuterVolumeSpecName: "kube-api-access-jrfzv") pod "7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2" (UID: "7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2"). InnerVolumeSpecName "kube-api-access-jrfzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.207844 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2" (UID: "7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.209904 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-inventory" (OuterVolumeSpecName: "inventory") pod "7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2" (UID: "7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.279697 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.279736 4715 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.279749 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrfzv\" (UniqueName: \"kubernetes.io/projected/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-kube-api-access-jrfzv\") on node \"crc\" DevicePath \"\"" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.279757 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.620618 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" event={"ID":"7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2","Type":"ContainerDied","Data":"9253df8af2b519313d9b816bb334933dcd87ba8d8dc1ea4f8480a96184aac6a8"} Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.621020 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9253df8af2b519313d9b816bb334933dcd87ba8d8dc1ea4f8480a96184aac6a8" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.621085 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.706128 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x"] Dec 10 10:00:33 crc kubenswrapper[4715]: E1210 10:00:33.706671 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerName="extract-utilities" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.706699 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerName="extract-utilities" Dec 10 10:00:33 crc kubenswrapper[4715]: E1210 10:00:33.706718 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb60d0d6-63d3-4453-9ae4-365ff52a48de" containerName="collect-profiles" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.706729 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb60d0d6-63d3-4453-9ae4-365ff52a48de" containerName="collect-profiles" Dec 10 10:00:33 crc kubenswrapper[4715]: E1210 10:00:33.706742 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerName="registry-server" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.706754 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerName="registry-server" Dec 10 10:00:33 crc kubenswrapper[4715]: E1210 10:00:33.706811 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerName="extract-content" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.706819 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerName="extract-content" Dec 10 10:00:33 crc kubenswrapper[4715]: E1210 10:00:33.706841 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.706853 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.707132 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb60d0d6-63d3-4453-9ae4-365ff52a48de" containerName="collect-profiles" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.707163 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.707184 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c9fe615-64aa-490f-b79d-63a032c0b219" containerName="registry-server" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.707986 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.714247 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.714264 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.714757 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.715332 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.730729 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x"] Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.789192 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.789356 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.789430 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg2nc\" (UniqueName: \"kubernetes.io/projected/6d7bc0b2-e15d-4627-b901-bb6399f7f636-kube-api-access-cg2nc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.891219 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.891287 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg2nc\" (UniqueName: \"kubernetes.io/projected/6d7bc0b2-e15d-4627-b901-bb6399f7f636-kube-api-access-cg2nc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.891438 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.901061 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.901113 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:33 crc kubenswrapper[4715]: I1210 10:00:33.910720 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg2nc\" (UniqueName: \"kubernetes.io/projected/6d7bc0b2-e15d-4627-b901-bb6399f7f636-kube-api-access-cg2nc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:34 crc kubenswrapper[4715]: I1210 10:00:34.039513 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:00:34 crc kubenswrapper[4715]: I1210 10:00:34.567378 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x"] Dec 10 10:00:34 crc kubenswrapper[4715]: I1210 10:00:34.631073 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" event={"ID":"6d7bc0b2-e15d-4627-b901-bb6399f7f636","Type":"ContainerStarted","Data":"d370378ce590447ecfc05bab048e57496a2a33aa09f7cf4665aac2c1f3c2d592"} Dec 10 10:00:35 crc kubenswrapper[4715]: I1210 10:00:35.644083 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" event={"ID":"6d7bc0b2-e15d-4627-b901-bb6399f7f636","Type":"ContainerStarted","Data":"33618601dbfedbeb600ec49fd0762c0609754790846fa04ebacb00cf8b0670cc"} Dec 10 10:00:35 crc kubenswrapper[4715]: I1210 10:00:35.672400 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" podStartSLOduration=2.225051731 podStartE2EDuration="2.672376251s" podCreationTimestamp="2025-12-10 10:00:33 +0000 UTC" firstStartedPulling="2025-12-10 10:00:34.572724718 +0000 UTC m=+1597.316270969" lastFinishedPulling="2025-12-10 10:00:35.020049238 +0000 UTC m=+1597.763595489" observedRunningTime="2025-12-10 10:00:35.662362003 +0000 UTC m=+1598.405908254" watchObservedRunningTime="2025-12-10 10:00:35.672376251 +0000 UTC m=+1598.415922502" Dec 10 10:00:47 crc kubenswrapper[4715]: I1210 10:00:47.713897 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:00:47 crc kubenswrapper[4715]: I1210 10:00:47.715668 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.158064 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29422681-fb8s5"] Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.162616 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.174874 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29422681-fb8s5"] Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.274432 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-fernet-keys\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.274878 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-combined-ca-bundle\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.275231 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-config-data\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.275643 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj4h6\" (UniqueName: \"kubernetes.io/projected/2992fa3e-b3d4-42da-845e-a019e8e23cf1-kube-api-access-nj4h6\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.376858 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-combined-ca-bundle\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.376959 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-config-data\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.377019 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj4h6\" (UniqueName: \"kubernetes.io/projected/2992fa3e-b3d4-42da-845e-a019e8e23cf1-kube-api-access-nj4h6\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.377053 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-fernet-keys\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.383518 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-fernet-keys\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.383739 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-config-data\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.384016 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-combined-ca-bundle\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.397736 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj4h6\" (UniqueName: \"kubernetes.io/projected/2992fa3e-b3d4-42da-845e-a019e8e23cf1-kube-api-access-nj4h6\") pod \"keystone-cron-29422681-fb8s5\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.482337 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:00 crc kubenswrapper[4715]: I1210 10:01:00.954745 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29422681-fb8s5"] Dec 10 10:01:01 crc kubenswrapper[4715]: I1210 10:01:01.897360 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29422681-fb8s5" event={"ID":"2992fa3e-b3d4-42da-845e-a019e8e23cf1","Type":"ContainerStarted","Data":"a2e26aa4a4d385e4b116559c67f5d911a4096b9578226ac537b9bc135e07d386"} Dec 10 10:01:01 crc kubenswrapper[4715]: I1210 10:01:01.897837 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29422681-fb8s5" event={"ID":"2992fa3e-b3d4-42da-845e-a019e8e23cf1","Type":"ContainerStarted","Data":"d1d80c1315e827acd5d27f1bb0c5248adab313d2a9aac30e43035d0e55f0141f"} Dec 10 10:01:01 crc kubenswrapper[4715]: I1210 10:01:01.921505 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29422681-fb8s5" podStartSLOduration=1.9214897899999999 podStartE2EDuration="1.92148979s" podCreationTimestamp="2025-12-10 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 10:01:01.915045894 +0000 UTC m=+1624.658592145" watchObservedRunningTime="2025-12-10 10:01:01.92148979 +0000 UTC m=+1624.665036041" Dec 10 10:01:03 crc kubenswrapper[4715]: I1210 10:01:03.943144 4715 generic.go:334] "Generic (PLEG): container finished" podID="2992fa3e-b3d4-42da-845e-a019e8e23cf1" containerID="a2e26aa4a4d385e4b116559c67f5d911a4096b9578226ac537b9bc135e07d386" exitCode=0 Dec 10 10:01:03 crc kubenswrapper[4715]: I1210 10:01:03.943226 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29422681-fb8s5" event={"ID":"2992fa3e-b3d4-42da-845e-a019e8e23cf1","Type":"ContainerDied","Data":"a2e26aa4a4d385e4b116559c67f5d911a4096b9578226ac537b9bc135e07d386"} Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.313941 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.477388 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-combined-ca-bundle\") pod \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.477810 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj4h6\" (UniqueName: \"kubernetes.io/projected/2992fa3e-b3d4-42da-845e-a019e8e23cf1-kube-api-access-nj4h6\") pod \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.478067 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-config-data\") pod \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.478171 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-fernet-keys\") pod \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\" (UID: \"2992fa3e-b3d4-42da-845e-a019e8e23cf1\") " Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.486261 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2992fa3e-b3d4-42da-845e-a019e8e23cf1-kube-api-access-nj4h6" (OuterVolumeSpecName: "kube-api-access-nj4h6") pod "2992fa3e-b3d4-42da-845e-a019e8e23cf1" (UID: "2992fa3e-b3d4-42da-845e-a019e8e23cf1"). InnerVolumeSpecName "kube-api-access-nj4h6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.488231 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2992fa3e-b3d4-42da-845e-a019e8e23cf1" (UID: "2992fa3e-b3d4-42da-845e-a019e8e23cf1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.514552 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2992fa3e-b3d4-42da-845e-a019e8e23cf1" (UID: "2992fa3e-b3d4-42da-845e-a019e8e23cf1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.539151 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-config-data" (OuterVolumeSpecName: "config-data") pod "2992fa3e-b3d4-42da-845e-a019e8e23cf1" (UID: "2992fa3e-b3d4-42da-845e-a019e8e23cf1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.581040 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.581146 4715 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.581159 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2992fa3e-b3d4-42da-845e-a019e8e23cf1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.581174 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj4h6\" (UniqueName: \"kubernetes.io/projected/2992fa3e-b3d4-42da-845e-a019e8e23cf1-kube-api-access-nj4h6\") on node \"crc\" DevicePath \"\"" Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.964949 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29422681-fb8s5" event={"ID":"2992fa3e-b3d4-42da-845e-a019e8e23cf1","Type":"ContainerDied","Data":"d1d80c1315e827acd5d27f1bb0c5248adab313d2a9aac30e43035d0e55f0141f"} Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.965021 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1d80c1315e827acd5d27f1bb0c5248adab313d2a9aac30e43035d0e55f0141f" Dec 10 10:01:05 crc kubenswrapper[4715]: I1210 10:01:05.965039 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29422681-fb8s5" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.602443 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sgzsp"] Dec 10 10:01:06 crc kubenswrapper[4715]: E1210 10:01:06.604097 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2992fa3e-b3d4-42da-845e-a019e8e23cf1" containerName="keystone-cron" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.604130 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2992fa3e-b3d4-42da-845e-a019e8e23cf1" containerName="keystone-cron" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.604490 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2992fa3e-b3d4-42da-845e-a019e8e23cf1" containerName="keystone-cron" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.608520 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.616408 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgzsp"] Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.703871 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtsrc\" (UniqueName: \"kubernetes.io/projected/bd81c0fd-bfa1-4530-850f-37c5c9684798-kube-api-access-wtsrc\") pod \"redhat-marketplace-sgzsp\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.703931 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-catalog-content\") pod \"redhat-marketplace-sgzsp\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.704119 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-utilities\") pod \"redhat-marketplace-sgzsp\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.805908 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtsrc\" (UniqueName: \"kubernetes.io/projected/bd81c0fd-bfa1-4530-850f-37c5c9684798-kube-api-access-wtsrc\") pod \"redhat-marketplace-sgzsp\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.806360 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-catalog-content\") pod \"redhat-marketplace-sgzsp\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.806785 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-catalog-content\") pod \"redhat-marketplace-sgzsp\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.807126 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-utilities\") pod \"redhat-marketplace-sgzsp\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.807160 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-utilities\") pod \"redhat-marketplace-sgzsp\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.828937 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtsrc\" (UniqueName: \"kubernetes.io/projected/bd81c0fd-bfa1-4530-850f-37c5c9684798-kube-api-access-wtsrc\") pod \"redhat-marketplace-sgzsp\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:06 crc kubenswrapper[4715]: I1210 10:01:06.944316 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:07 crc kubenswrapper[4715]: I1210 10:01:07.446659 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgzsp"] Dec 10 10:01:08 crc kubenswrapper[4715]: I1210 10:01:08.002489 4715 generic.go:334] "Generic (PLEG): container finished" podID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerID="87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc" exitCode=0 Dec 10 10:01:08 crc kubenswrapper[4715]: I1210 10:01:08.003762 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgzsp" event={"ID":"bd81c0fd-bfa1-4530-850f-37c5c9684798","Type":"ContainerDied","Data":"87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc"} Dec 10 10:01:08 crc kubenswrapper[4715]: I1210 10:01:08.004077 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgzsp" event={"ID":"bd81c0fd-bfa1-4530-850f-37c5c9684798","Type":"ContainerStarted","Data":"f5214626a9796da638d18beee4f64da2aa6d85f08a1fe0553af5dfef75fc1c1d"} Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.053326 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-7sk54"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.063086 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6793-account-create-update-6f4cp"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.074417 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-lm8rz"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.084599 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f4e8-account-create-update-xbb2q"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.093264 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6121-account-create-update-zw9vn"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.100907 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-ktxnt"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.108646 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6793-account-create-update-6f4cp"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.116377 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-7sk54"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.123344 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-lm8rz"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.130646 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-ktxnt"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.137998 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f4e8-account-create-update-xbb2q"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.146601 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6121-account-create-update-zw9vn"] Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.616226 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d135565-25c1-467f-9556-c5114c704da0" path="/var/lib/kubelet/pods/0d135565-25c1-467f-9556-c5114c704da0/volumes" Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.617774 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44adefd5-a106-405e-b70c-464d3d111dfd" path="/var/lib/kubelet/pods/44adefd5-a106-405e-b70c-464d3d111dfd/volumes" Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.618646 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9291a4d5-3925-428d-9351-30e4bc4186c8" path="/var/lib/kubelet/pods/9291a4d5-3925-428d-9351-30e4bc4186c8/volumes" Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.619445 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd6bdd7f-e4fd-4c76-8f2e-286491ddb350" path="/var/lib/kubelet/pods/bd6bdd7f-e4fd-4c76-8f2e-286491ddb350/volumes" Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.620705 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e10b59-95b4-4a8c-b69d-392afceeb5cb" path="/var/lib/kubelet/pods/d3e10b59-95b4-4a8c-b69d-392afceeb5cb/volumes" Dec 10 10:01:09 crc kubenswrapper[4715]: I1210 10:01:09.621323 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7026670-8991-4a8f-b70d-3f3e6291fc6e" path="/var/lib/kubelet/pods/f7026670-8991-4a8f-b70d-3f3e6291fc6e/volumes" Dec 10 10:01:10 crc kubenswrapper[4715]: I1210 10:01:10.036101 4715 generic.go:334] "Generic (PLEG): container finished" podID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerID="f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d" exitCode=0 Dec 10 10:01:10 crc kubenswrapper[4715]: I1210 10:01:10.036150 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgzsp" event={"ID":"bd81c0fd-bfa1-4530-850f-37c5c9684798","Type":"ContainerDied","Data":"f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d"} Dec 10 10:01:11 crc kubenswrapper[4715]: I1210 10:01:11.047606 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgzsp" event={"ID":"bd81c0fd-bfa1-4530-850f-37c5c9684798","Type":"ContainerStarted","Data":"59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9"} Dec 10 10:01:11 crc kubenswrapper[4715]: I1210 10:01:11.075313 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sgzsp" podStartSLOduration=2.372305963 podStartE2EDuration="5.075291582s" podCreationTimestamp="2025-12-10 10:01:06 +0000 UTC" firstStartedPulling="2025-12-10 10:01:08.007657353 +0000 UTC m=+1630.751203624" lastFinishedPulling="2025-12-10 10:01:10.710642992 +0000 UTC m=+1633.454189243" observedRunningTime="2025-12-10 10:01:11.06551641 +0000 UTC m=+1633.809062671" watchObservedRunningTime="2025-12-10 10:01:11.075291582 +0000 UTC m=+1633.818837833" Dec 10 10:01:16 crc kubenswrapper[4715]: I1210 10:01:16.945102 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:16 crc kubenswrapper[4715]: I1210 10:01:16.945804 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:17 crc kubenswrapper[4715]: I1210 10:01:17.004872 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:17 crc kubenswrapper[4715]: I1210 10:01:17.153576 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:17 crc kubenswrapper[4715]: I1210 10:01:17.245726 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgzsp"] Dec 10 10:01:17 crc kubenswrapper[4715]: I1210 10:01:17.714585 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:01:17 crc kubenswrapper[4715]: I1210 10:01:17.714666 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:01:17 crc kubenswrapper[4715]: I1210 10:01:17.714724 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 10:01:17 crc kubenswrapper[4715]: I1210 10:01:17.715610 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 10:01:17 crc kubenswrapper[4715]: I1210 10:01:17.715757 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" gracePeriod=600 Dec 10 10:01:17 crc kubenswrapper[4715]: E1210 10:01:17.864225 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:01:18 crc kubenswrapper[4715]: I1210 10:01:18.118596 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" exitCode=0 Dec 10 10:01:18 crc kubenswrapper[4715]: I1210 10:01:18.118661 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7"} Dec 10 10:01:18 crc kubenswrapper[4715]: I1210 10:01:18.119059 4715 scope.go:117] "RemoveContainer" containerID="43f0315f5b714d86cf23ed6f6c468531cbdb76f130137362d46c5a9fc1556442" Dec 10 10:01:18 crc kubenswrapper[4715]: I1210 10:01:18.119959 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:01:18 crc kubenswrapper[4715]: E1210 10:01:18.120680 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.129346 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sgzsp" podUID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerName="registry-server" containerID="cri-o://59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9" gracePeriod=2 Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.610624 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.772707 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtsrc\" (UniqueName: \"kubernetes.io/projected/bd81c0fd-bfa1-4530-850f-37c5c9684798-kube-api-access-wtsrc\") pod \"bd81c0fd-bfa1-4530-850f-37c5c9684798\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.774000 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-catalog-content\") pod \"bd81c0fd-bfa1-4530-850f-37c5c9684798\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.779225 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd81c0fd-bfa1-4530-850f-37c5c9684798-kube-api-access-wtsrc" (OuterVolumeSpecName: "kube-api-access-wtsrc") pod "bd81c0fd-bfa1-4530-850f-37c5c9684798" (UID: "bd81c0fd-bfa1-4530-850f-37c5c9684798"). InnerVolumeSpecName "kube-api-access-wtsrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.779514 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-utilities\") pod \"bd81c0fd-bfa1-4530-850f-37c5c9684798\" (UID: \"bd81c0fd-bfa1-4530-850f-37c5c9684798\") " Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.780462 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-utilities" (OuterVolumeSpecName: "utilities") pod "bd81c0fd-bfa1-4530-850f-37c5c9684798" (UID: "bd81c0fd-bfa1-4530-850f-37c5c9684798"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.784424 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.784473 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtsrc\" (UniqueName: \"kubernetes.io/projected/bd81c0fd-bfa1-4530-850f-37c5c9684798-kube-api-access-wtsrc\") on node \"crc\" DevicePath \"\"" Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.802165 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd81c0fd-bfa1-4530-850f-37c5c9684798" (UID: "bd81c0fd-bfa1-4530-850f-37c5c9684798"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:01:19 crc kubenswrapper[4715]: I1210 10:01:19.886085 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd81c0fd-bfa1-4530-850f-37c5c9684798-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.147175 4715 generic.go:334] "Generic (PLEG): container finished" podID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerID="59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9" exitCode=0 Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.147198 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sgzsp" Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.147221 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgzsp" event={"ID":"bd81c0fd-bfa1-4530-850f-37c5c9684798","Type":"ContainerDied","Data":"59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9"} Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.147272 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sgzsp" event={"ID":"bd81c0fd-bfa1-4530-850f-37c5c9684798","Type":"ContainerDied","Data":"f5214626a9796da638d18beee4f64da2aa6d85f08a1fe0553af5dfef75fc1c1d"} Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.147293 4715 scope.go:117] "RemoveContainer" containerID="59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9" Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.177648 4715 scope.go:117] "RemoveContainer" containerID="f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d" Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.182842 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgzsp"] Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.190767 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sgzsp"] Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.213452 4715 scope.go:117] "RemoveContainer" containerID="87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc" Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.243200 4715 scope.go:117] "RemoveContainer" containerID="59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9" Dec 10 10:01:20 crc kubenswrapper[4715]: E1210 10:01:20.243578 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9\": container with ID starting with 59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9 not found: ID does not exist" containerID="59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9" Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.243621 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9"} err="failed to get container status \"59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9\": rpc error: code = NotFound desc = could not find container \"59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9\": container with ID starting with 59a196dd9647325e92856714144e9e1c60159873691a71534100ac735d5accb9 not found: ID does not exist" Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.243654 4715 scope.go:117] "RemoveContainer" containerID="f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d" Dec 10 10:01:20 crc kubenswrapper[4715]: E1210 10:01:20.244382 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d\": container with ID starting with f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d not found: ID does not exist" containerID="f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d" Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.244400 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d"} err="failed to get container status \"f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d\": rpc error: code = NotFound desc = could not find container \"f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d\": container with ID starting with f27c061dc2f3838d43e783b2bb4b10a0cfaad45933b7cb80ff93550bbec5ca6d not found: ID does not exist" Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.244417 4715 scope.go:117] "RemoveContainer" containerID="87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc" Dec 10 10:01:20 crc kubenswrapper[4715]: E1210 10:01:20.244596 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc\": container with ID starting with 87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc not found: ID does not exist" containerID="87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc" Dec 10 10:01:20 crc kubenswrapper[4715]: I1210 10:01:20.244622 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc"} err="failed to get container status \"87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc\": rpc error: code = NotFound desc = could not find container \"87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc\": container with ID starting with 87c62900c9dfbc8a09d039d7d35bdc3b32f6f6054da1c9733534f3249c14dedc not found: ID does not exist" Dec 10 10:01:21 crc kubenswrapper[4715]: I1210 10:01:21.616183 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd81c0fd-bfa1-4530-850f-37c5c9684798" path="/var/lib/kubelet/pods/bd81c0fd-bfa1-4530-850f-37c5c9684798/volumes" Dec 10 10:01:21 crc kubenswrapper[4715]: I1210 10:01:21.715543 4715 scope.go:117] "RemoveContainer" containerID="b4a38ed4e28f2aa3b65206788a62369dbc91ece4d8cc1f1f988e5c7347996839" Dec 10 10:01:21 crc kubenswrapper[4715]: I1210 10:01:21.755410 4715 scope.go:117] "RemoveContainer" containerID="ec4511826d538604be5f4658736ea25f1e9ea24edcf9cad47ecd7104d46c4ad9" Dec 10 10:01:21 crc kubenswrapper[4715]: I1210 10:01:21.786294 4715 scope.go:117] "RemoveContainer" containerID="ee91540b6671e2218402316ba4ab4386d6cb12134c441b70169d8512b75cd721" Dec 10 10:01:21 crc kubenswrapper[4715]: I1210 10:01:21.837072 4715 scope.go:117] "RemoveContainer" containerID="0ff053fbe0e06f75d30846932e658aa9b03b2b4915164ce2d48cfca95161f99e" Dec 10 10:01:21 crc kubenswrapper[4715]: I1210 10:01:21.886691 4715 scope.go:117] "RemoveContainer" containerID="021b0cdf2963aa3d9ec5796b1f3bcd52205509f295c4c57361cd9d404a8693a5" Dec 10 10:01:21 crc kubenswrapper[4715]: I1210 10:01:21.942326 4715 scope.go:117] "RemoveContainer" containerID="9134f9fb25c8bc74c9f5257182933ef4bff717392eda7190418d410117036b94" Dec 10 10:01:28 crc kubenswrapper[4715]: I1210 10:01:28.605014 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:01:28 crc kubenswrapper[4715]: E1210 10:01:28.605835 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:01:36 crc kubenswrapper[4715]: I1210 10:01:36.079305 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-6h5ss"] Dec 10 10:01:36 crc kubenswrapper[4715]: I1210 10:01:36.095940 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5098-account-create-update-n88lg"] Dec 10 10:01:36 crc kubenswrapper[4715]: I1210 10:01:36.112835 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-6h5ss"] Dec 10 10:01:36 crc kubenswrapper[4715]: I1210 10:01:36.124514 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5098-account-create-update-n88lg"] Dec 10 10:01:36 crc kubenswrapper[4715]: I1210 10:01:36.136333 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-68dkb"] Dec 10 10:01:36 crc kubenswrapper[4715]: I1210 10:01:36.147760 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-x26bs"] Dec 10 10:01:36 crc kubenswrapper[4715]: I1210 10:01:36.159091 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-x26bs"] Dec 10 10:01:36 crc kubenswrapper[4715]: I1210 10:01:36.171133 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-68dkb"] Dec 10 10:01:37 crc kubenswrapper[4715]: I1210 10:01:37.618669 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b5f1b4-0e7f-4b10-808e-845621bc68d2" path="/var/lib/kubelet/pods/68b5f1b4-0e7f-4b10-808e-845621bc68d2/volumes" Dec 10 10:01:37 crc kubenswrapper[4715]: I1210 10:01:37.619455 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0f80a84-53e6-4c55-89a0-fb526f3998c1" path="/var/lib/kubelet/pods/b0f80a84-53e6-4c55-89a0-fb526f3998c1/volumes" Dec 10 10:01:37 crc kubenswrapper[4715]: I1210 10:01:37.620359 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2228f25-b904-46ba-8999-23071a80094c" path="/var/lib/kubelet/pods/b2228f25-b904-46ba-8999-23071a80094c/volumes" Dec 10 10:01:37 crc kubenswrapper[4715]: I1210 10:01:37.620943 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cae7741b-9dd1-459a-a05e-c61a3c42b782" path="/var/lib/kubelet/pods/cae7741b-9dd1-459a-a05e-c61a3c42b782/volumes" Dec 10 10:01:39 crc kubenswrapper[4715]: I1210 10:01:39.046589 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-0bb5-account-create-update-7gr2k"] Dec 10 10:01:39 crc kubenswrapper[4715]: I1210 10:01:39.062287 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-38cd-account-create-update-pgrt5"] Dec 10 10:01:39 crc kubenswrapper[4715]: I1210 10:01:39.074581 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-0bb5-account-create-update-7gr2k"] Dec 10 10:01:39 crc kubenswrapper[4715]: I1210 10:01:39.084595 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-38cd-account-create-update-pgrt5"] Dec 10 10:01:39 crc kubenswrapper[4715]: I1210 10:01:39.617224 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6360fe51-a8bf-4b95-9d03-163543c7363e" path="/var/lib/kubelet/pods/6360fe51-a8bf-4b95-9d03-163543c7363e/volumes" Dec 10 10:01:39 crc kubenswrapper[4715]: I1210 10:01:39.618071 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf" path="/var/lib/kubelet/pods/6cfe62cd-f6dc-400d-b6cf-3484a6d5eacf/volumes" Dec 10 10:01:40 crc kubenswrapper[4715]: I1210 10:01:40.605263 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:01:40 crc kubenswrapper[4715]: E1210 10:01:40.605863 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:01:43 crc kubenswrapper[4715]: I1210 10:01:43.033822 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-gd24z"] Dec 10 10:01:43 crc kubenswrapper[4715]: I1210 10:01:43.042740 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-gd24z"] Dec 10 10:01:43 crc kubenswrapper[4715]: I1210 10:01:43.626632 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad144d0-9a73-4736-8f88-43dd295478d7" path="/var/lib/kubelet/pods/8ad144d0-9a73-4736-8f88-43dd295478d7/volumes" Dec 10 10:01:45 crc kubenswrapper[4715]: I1210 10:01:45.046123 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-bhnvd"] Dec 10 10:01:45 crc kubenswrapper[4715]: I1210 10:01:45.061817 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-bhnvd"] Dec 10 10:01:45 crc kubenswrapper[4715]: I1210 10:01:45.618873 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="108964f4-918d-4a16-93d2-85f0ee798efc" path="/var/lib/kubelet/pods/108964f4-918d-4a16-93d2-85f0ee798efc/volumes" Dec 10 10:01:55 crc kubenswrapper[4715]: I1210 10:01:55.605133 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:01:55 crc kubenswrapper[4715]: E1210 10:01:55.606194 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:02:06 crc kubenswrapper[4715]: I1210 10:02:06.605195 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:02:06 crc kubenswrapper[4715]: E1210 10:02:06.605929 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:02:11 crc kubenswrapper[4715]: I1210 10:02:11.727314 4715 generic.go:334] "Generic (PLEG): container finished" podID="6d7bc0b2-e15d-4627-b901-bb6399f7f636" containerID="33618601dbfedbeb600ec49fd0762c0609754790846fa04ebacb00cf8b0670cc" exitCode=0 Dec 10 10:02:11 crc kubenswrapper[4715]: I1210 10:02:11.727407 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" event={"ID":"6d7bc0b2-e15d-4627-b901-bb6399f7f636","Type":"ContainerDied","Data":"33618601dbfedbeb600ec49fd0762c0609754790846fa04ebacb00cf8b0670cc"} Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.099988 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.217391 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg2nc\" (UniqueName: \"kubernetes.io/projected/6d7bc0b2-e15d-4627-b901-bb6399f7f636-kube-api-access-cg2nc\") pod \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.217503 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-ssh-key\") pod \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.217564 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-inventory\") pod \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\" (UID: \"6d7bc0b2-e15d-4627-b901-bb6399f7f636\") " Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.237151 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d7bc0b2-e15d-4627-b901-bb6399f7f636-kube-api-access-cg2nc" (OuterVolumeSpecName: "kube-api-access-cg2nc") pod "6d7bc0b2-e15d-4627-b901-bb6399f7f636" (UID: "6d7bc0b2-e15d-4627-b901-bb6399f7f636"). InnerVolumeSpecName "kube-api-access-cg2nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.295656 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-inventory" (OuterVolumeSpecName: "inventory") pod "6d7bc0b2-e15d-4627-b901-bb6399f7f636" (UID: "6d7bc0b2-e15d-4627-b901-bb6399f7f636"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.299898 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6d7bc0b2-e15d-4627-b901-bb6399f7f636" (UID: "6d7bc0b2-e15d-4627-b901-bb6399f7f636"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.319610 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg2nc\" (UniqueName: \"kubernetes.io/projected/6d7bc0b2-e15d-4627-b901-bb6399f7f636-kube-api-access-cg2nc\") on node \"crc\" DevicePath \"\"" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.319653 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.319665 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d7bc0b2-e15d-4627-b901-bb6399f7f636-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.746333 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" event={"ID":"6d7bc0b2-e15d-4627-b901-bb6399f7f636","Type":"ContainerDied","Data":"d370378ce590447ecfc05bab048e57496a2a33aa09f7cf4665aac2c1f3c2d592"} Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.746377 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d370378ce590447ecfc05bab048e57496a2a33aa09f7cf4665aac2c1f3c2d592" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.746399 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.837020 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl"] Dec 10 10:02:13 crc kubenswrapper[4715]: E1210 10:02:13.837462 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerName="extract-utilities" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.837475 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerName="extract-utilities" Dec 10 10:02:13 crc kubenswrapper[4715]: E1210 10:02:13.837523 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerName="registry-server" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.837529 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerName="registry-server" Dec 10 10:02:13 crc kubenswrapper[4715]: E1210 10:02:13.837536 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d7bc0b2-e15d-4627-b901-bb6399f7f636" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.837543 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d7bc0b2-e15d-4627-b901-bb6399f7f636" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 10:02:13 crc kubenswrapper[4715]: E1210 10:02:13.837554 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerName="extract-content" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.837560 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerName="extract-content" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.837732 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd81c0fd-bfa1-4530-850f-37c5c9684798" containerName="registry-server" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.837759 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d7bc0b2-e15d-4627-b901-bb6399f7f636" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.838549 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.842401 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.842637 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.842777 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.842959 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.852498 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl"] Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.980836 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.981150 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:13 crc kubenswrapper[4715]: I1210 10:02:13.981223 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh6kr\" (UniqueName: \"kubernetes.io/projected/42754cce-1e7d-416d-a1fa-118ae307880f-kube-api-access-xh6kr\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:14 crc kubenswrapper[4715]: I1210 10:02:14.082873 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:14 crc kubenswrapper[4715]: I1210 10:02:14.082941 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:14 crc kubenswrapper[4715]: I1210 10:02:14.083042 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh6kr\" (UniqueName: \"kubernetes.io/projected/42754cce-1e7d-416d-a1fa-118ae307880f-kube-api-access-xh6kr\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:14 crc kubenswrapper[4715]: I1210 10:02:14.087185 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:14 crc kubenswrapper[4715]: I1210 10:02:14.087951 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:14 crc kubenswrapper[4715]: I1210 10:02:14.102675 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh6kr\" (UniqueName: \"kubernetes.io/projected/42754cce-1e7d-416d-a1fa-118ae307880f-kube-api-access-xh6kr\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:14 crc kubenswrapper[4715]: I1210 10:02:14.156579 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:02:14 crc kubenswrapper[4715]: I1210 10:02:14.718678 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl"] Dec 10 10:02:14 crc kubenswrapper[4715]: I1210 10:02:14.757724 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" event={"ID":"42754cce-1e7d-416d-a1fa-118ae307880f","Type":"ContainerStarted","Data":"f44e5f8ec97b4a01e5f3731d3769385813b005383bcf6db6a14f8abad2e95c42"} Dec 10 10:02:15 crc kubenswrapper[4715]: I1210 10:02:15.767390 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" event={"ID":"42754cce-1e7d-416d-a1fa-118ae307880f","Type":"ContainerStarted","Data":"57fb13097d7d9ff6745a4324207689c41beea2e176610c911ef1bcf340cf3c24"} Dec 10 10:02:15 crc kubenswrapper[4715]: I1210 10:02:15.787446 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" podStartSLOduration=2.261929437 podStartE2EDuration="2.787422411s" podCreationTimestamp="2025-12-10 10:02:13 +0000 UTC" firstStartedPulling="2025-12-10 10:02:14.744644097 +0000 UTC m=+1697.488190338" lastFinishedPulling="2025-12-10 10:02:15.270137061 +0000 UTC m=+1698.013683312" observedRunningTime="2025-12-10 10:02:15.781150332 +0000 UTC m=+1698.524696583" watchObservedRunningTime="2025-12-10 10:02:15.787422411 +0000 UTC m=+1698.530968662" Dec 10 10:02:18 crc kubenswrapper[4715]: I1210 10:02:18.066345 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2vj6t"] Dec 10 10:02:18 crc kubenswrapper[4715]: I1210 10:02:18.084559 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2vj6t"] Dec 10 10:02:19 crc kubenswrapper[4715]: I1210 10:02:19.605847 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:02:19 crc kubenswrapper[4715]: E1210 10:02:19.606636 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:02:19 crc kubenswrapper[4715]: I1210 10:02:19.621022 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a08f4204-23b4-425a-a116-cfd37741c87f" path="/var/lib/kubelet/pods/a08f4204-23b4-425a-a116-cfd37741c87f/volumes" Dec 10 10:02:22 crc kubenswrapper[4715]: I1210 10:02:22.084469 4715 scope.go:117] "RemoveContainer" containerID="9e96cdcfdabf254db79354a432fcf5f570a5edf6123cbce6e6057b52ace584df" Dec 10 10:02:22 crc kubenswrapper[4715]: I1210 10:02:22.114216 4715 scope.go:117] "RemoveContainer" containerID="6ec18426bc83ba73a7fa8c4317f0d027df1e768fd368168af3230338213e4d0c" Dec 10 10:02:22 crc kubenswrapper[4715]: I1210 10:02:22.169135 4715 scope.go:117] "RemoveContainer" containerID="2405ad40366ebf0445c1b61deba023d87a8460937c11c891e3a3cb9a30b1b503" Dec 10 10:02:22 crc kubenswrapper[4715]: I1210 10:02:22.215306 4715 scope.go:117] "RemoveContainer" containerID="0c4f3f531995496685f907c8d2b13f8583be7b428dd7c227e2622600162bb714" Dec 10 10:02:22 crc kubenswrapper[4715]: I1210 10:02:22.269166 4715 scope.go:117] "RemoveContainer" containerID="6292a5055b87b5d45f479cf639b29d193fb29a445d5e3d584134842a70af9e8a" Dec 10 10:02:22 crc kubenswrapper[4715]: I1210 10:02:22.291169 4715 scope.go:117] "RemoveContainer" containerID="c994c59c3996db0033fcbfd436c5edf439427180f8eb91692bfb34864bf118b9" Dec 10 10:02:22 crc kubenswrapper[4715]: I1210 10:02:22.337693 4715 scope.go:117] "RemoveContainer" containerID="e68e45416c5c46e48d36f198c2ea5aa1081bf6d0dbc2a1e9a18ad82f2c64f29f" Dec 10 10:02:22 crc kubenswrapper[4715]: I1210 10:02:22.367531 4715 scope.go:117] "RemoveContainer" containerID="f5f1356f1edc87ef4de08ea55fcefe3e1ec816f67580dd41112ccc4c4ccab24e" Dec 10 10:02:22 crc kubenswrapper[4715]: I1210 10:02:22.386255 4715 scope.go:117] "RemoveContainer" containerID="2e9be1f8f85f577f3901b5548949f35e2fe59a0f0b9e968f924ae21ca3f853fb" Dec 10 10:02:25 crc kubenswrapper[4715]: I1210 10:02:25.043348 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-zq2xt"] Dec 10 10:02:25 crc kubenswrapper[4715]: I1210 10:02:25.063960 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-zq2xt"] Dec 10 10:02:25 crc kubenswrapper[4715]: I1210 10:02:25.615159 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53507dfd-6f17-486f-8770-c073ec1b42a3" path="/var/lib/kubelet/pods/53507dfd-6f17-486f-8770-c073ec1b42a3/volumes" Dec 10 10:02:32 crc kubenswrapper[4715]: I1210 10:02:32.604644 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:02:32 crc kubenswrapper[4715]: E1210 10:02:32.605356 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:02:33 crc kubenswrapper[4715]: I1210 10:02:33.038962 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-x2v2z"] Dec 10 10:02:33 crc kubenswrapper[4715]: I1210 10:02:33.049714 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-x2v2z"] Dec 10 10:02:33 crc kubenswrapper[4715]: I1210 10:02:33.616053 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27133872-3a45-45a6-9d91-f645c0c9b264" path="/var/lib/kubelet/pods/27133872-3a45-45a6-9d91-f645c0c9b264/volumes" Dec 10 10:02:42 crc kubenswrapper[4715]: I1210 10:02:42.042886 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-gzr48"] Dec 10 10:02:42 crc kubenswrapper[4715]: I1210 10:02:42.058294 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-gzr48"] Dec 10 10:02:43 crc kubenswrapper[4715]: I1210 10:02:43.045885 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-28hsp"] Dec 10 10:02:43 crc kubenswrapper[4715]: I1210 10:02:43.055850 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-28hsp"] Dec 10 10:02:43 crc kubenswrapper[4715]: I1210 10:02:43.617123 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66856339-f4ab-4102-93a8-adc952877ccd" path="/var/lib/kubelet/pods/66856339-f4ab-4102-93a8-adc952877ccd/volumes" Dec 10 10:02:43 crc kubenswrapper[4715]: I1210 10:02:43.622696 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf72f3c-8033-4018-b330-72fa2774f402" path="/var/lib/kubelet/pods/8bf72f3c-8033-4018-b330-72fa2774f402/volumes" Dec 10 10:02:45 crc kubenswrapper[4715]: I1210 10:02:45.604749 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:02:45 crc kubenswrapper[4715]: E1210 10:02:45.605072 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.688789 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b65bt"] Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.691424 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.707585 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b65bt"] Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.798732 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-catalog-content\") pod \"redhat-operators-b65bt\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.798801 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp6hc\" (UniqueName: \"kubernetes.io/projected/ad17376e-106a-4aed-abd5-bbc9a9737b11-kube-api-access-wp6hc\") pod \"redhat-operators-b65bt\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.799950 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-utilities\") pod \"redhat-operators-b65bt\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.902675 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-catalog-content\") pod \"redhat-operators-b65bt\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.902747 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp6hc\" (UniqueName: \"kubernetes.io/projected/ad17376e-106a-4aed-abd5-bbc9a9737b11-kube-api-access-wp6hc\") pod \"redhat-operators-b65bt\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.902807 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-utilities\") pod \"redhat-operators-b65bt\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.903323 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-catalog-content\") pod \"redhat-operators-b65bt\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.903467 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-utilities\") pod \"redhat-operators-b65bt\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:52 crc kubenswrapper[4715]: I1210 10:02:52.930336 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp6hc\" (UniqueName: \"kubernetes.io/projected/ad17376e-106a-4aed-abd5-bbc9a9737b11-kube-api-access-wp6hc\") pod \"redhat-operators-b65bt\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:53 crc kubenswrapper[4715]: I1210 10:02:53.014780 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:02:53 crc kubenswrapper[4715]: I1210 10:02:53.559633 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b65bt"] Dec 10 10:02:54 crc kubenswrapper[4715]: I1210 10:02:54.134130 4715 generic.go:334] "Generic (PLEG): container finished" podID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerID="0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741" exitCode=0 Dec 10 10:02:54 crc kubenswrapper[4715]: I1210 10:02:54.134267 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b65bt" event={"ID":"ad17376e-106a-4aed-abd5-bbc9a9737b11","Type":"ContainerDied","Data":"0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741"} Dec 10 10:02:54 crc kubenswrapper[4715]: I1210 10:02:54.134512 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b65bt" event={"ID":"ad17376e-106a-4aed-abd5-bbc9a9737b11","Type":"ContainerStarted","Data":"54a9153285bc9a0b412a6eed1b0d02c7292e09f393f873ba8d4b318e957b60c6"} Dec 10 10:02:55 crc kubenswrapper[4715]: I1210 10:02:55.172540 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b65bt" event={"ID":"ad17376e-106a-4aed-abd5-bbc9a9737b11","Type":"ContainerStarted","Data":"ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1"} Dec 10 10:02:58 crc kubenswrapper[4715]: I1210 10:02:58.209046 4715 generic.go:334] "Generic (PLEG): container finished" podID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerID="ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1" exitCode=0 Dec 10 10:02:58 crc kubenswrapper[4715]: I1210 10:02:58.209191 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b65bt" event={"ID":"ad17376e-106a-4aed-abd5-bbc9a9737b11","Type":"ContainerDied","Data":"ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1"} Dec 10 10:02:59 crc kubenswrapper[4715]: I1210 10:02:59.222578 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b65bt" event={"ID":"ad17376e-106a-4aed-abd5-bbc9a9737b11","Type":"ContainerStarted","Data":"0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7"} Dec 10 10:02:59 crc kubenswrapper[4715]: I1210 10:02:59.244423 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b65bt" podStartSLOduration=2.547815549 podStartE2EDuration="7.244405167s" podCreationTimestamp="2025-12-10 10:02:52 +0000 UTC" firstStartedPulling="2025-12-10 10:02:54.135738757 +0000 UTC m=+1736.879285008" lastFinishedPulling="2025-12-10 10:02:58.832328375 +0000 UTC m=+1741.575874626" observedRunningTime="2025-12-10 10:02:59.239780425 +0000 UTC m=+1741.983326676" watchObservedRunningTime="2025-12-10 10:02:59.244405167 +0000 UTC m=+1741.987951418" Dec 10 10:02:59 crc kubenswrapper[4715]: I1210 10:02:59.604761 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:02:59 crc kubenswrapper[4715]: E1210 10:02:59.605118 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:03:03 crc kubenswrapper[4715]: I1210 10:03:03.015819 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:03:03 crc kubenswrapper[4715]: I1210 10:03:03.016407 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:03:04 crc kubenswrapper[4715]: I1210 10:03:04.062707 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b65bt" podUID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerName="registry-server" probeResult="failure" output=< Dec 10 10:03:04 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 10 10:03:04 crc kubenswrapper[4715]: > Dec 10 10:03:13 crc kubenswrapper[4715]: I1210 10:03:13.075587 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:03:13 crc kubenswrapper[4715]: I1210 10:03:13.133841 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:03:13 crc kubenswrapper[4715]: I1210 10:03:13.328042 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b65bt"] Dec 10 10:03:13 crc kubenswrapper[4715]: I1210 10:03:13.606077 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:03:13 crc kubenswrapper[4715]: E1210 10:03:13.606431 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:03:14 crc kubenswrapper[4715]: I1210 10:03:14.369998 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b65bt" podUID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerName="registry-server" containerID="cri-o://0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7" gracePeriod=2 Dec 10 10:03:14 crc kubenswrapper[4715]: I1210 10:03:14.918388 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.106518 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-utilities\") pod \"ad17376e-106a-4aed-abd5-bbc9a9737b11\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.106779 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-catalog-content\") pod \"ad17376e-106a-4aed-abd5-bbc9a9737b11\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.106814 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp6hc\" (UniqueName: \"kubernetes.io/projected/ad17376e-106a-4aed-abd5-bbc9a9737b11-kube-api-access-wp6hc\") pod \"ad17376e-106a-4aed-abd5-bbc9a9737b11\" (UID: \"ad17376e-106a-4aed-abd5-bbc9a9737b11\") " Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.107348 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-utilities" (OuterVolumeSpecName: "utilities") pod "ad17376e-106a-4aed-abd5-bbc9a9737b11" (UID: "ad17376e-106a-4aed-abd5-bbc9a9737b11"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.126621 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad17376e-106a-4aed-abd5-bbc9a9737b11-kube-api-access-wp6hc" (OuterVolumeSpecName: "kube-api-access-wp6hc") pod "ad17376e-106a-4aed-abd5-bbc9a9737b11" (UID: "ad17376e-106a-4aed-abd5-bbc9a9737b11"). InnerVolumeSpecName "kube-api-access-wp6hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.209071 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp6hc\" (UniqueName: \"kubernetes.io/projected/ad17376e-106a-4aed-abd5-bbc9a9737b11-kube-api-access-wp6hc\") on node \"crc\" DevicePath \"\"" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.209113 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.223568 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad17376e-106a-4aed-abd5-bbc9a9737b11" (UID: "ad17376e-106a-4aed-abd5-bbc9a9737b11"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.310857 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad17376e-106a-4aed-abd5-bbc9a9737b11-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.381475 4715 generic.go:334] "Generic (PLEG): container finished" podID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerID="0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7" exitCode=0 Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.381523 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b65bt" event={"ID":"ad17376e-106a-4aed-abd5-bbc9a9737b11","Type":"ContainerDied","Data":"0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7"} Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.381550 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b65bt" event={"ID":"ad17376e-106a-4aed-abd5-bbc9a9737b11","Type":"ContainerDied","Data":"54a9153285bc9a0b412a6eed1b0d02c7292e09f393f873ba8d4b318e957b60c6"} Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.381566 4715 scope.go:117] "RemoveContainer" containerID="0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.381697 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b65bt" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.408417 4715 scope.go:117] "RemoveContainer" containerID="ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.423064 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b65bt"] Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.432696 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b65bt"] Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.440746 4715 scope.go:117] "RemoveContainer" containerID="0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.508054 4715 scope.go:117] "RemoveContainer" containerID="0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7" Dec 10 10:03:15 crc kubenswrapper[4715]: E1210 10:03:15.510116 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7\": container with ID starting with 0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7 not found: ID does not exist" containerID="0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.510173 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7"} err="failed to get container status \"0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7\": rpc error: code = NotFound desc = could not find container \"0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7\": container with ID starting with 0f96a42e35736afb3e5cf5adeebce50f9f83204c811c978bc05e56c8d9754ff7 not found: ID does not exist" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.510207 4715 scope.go:117] "RemoveContainer" containerID="ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1" Dec 10 10:03:15 crc kubenswrapper[4715]: E1210 10:03:15.510822 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1\": container with ID starting with ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1 not found: ID does not exist" containerID="ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.510877 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1"} err="failed to get container status \"ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1\": rpc error: code = NotFound desc = could not find container \"ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1\": container with ID starting with ae915c8bb9df7de90101066c5a4400a6b2d048ebcc52c4241be829c151fe84b1 not found: ID does not exist" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.511030 4715 scope.go:117] "RemoveContainer" containerID="0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741" Dec 10 10:03:15 crc kubenswrapper[4715]: E1210 10:03:15.511388 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741\": container with ID starting with 0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741 not found: ID does not exist" containerID="0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.511424 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741"} err="failed to get container status \"0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741\": rpc error: code = NotFound desc = could not find container \"0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741\": container with ID starting with 0b6e02c1a1392b9b55cc9bb819dff95d2c13ff8c5ec00326c0ebdc4c91003741 not found: ID does not exist" Dec 10 10:03:15 crc kubenswrapper[4715]: I1210 10:03:15.616849 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad17376e-106a-4aed-abd5-bbc9a9737b11" path="/var/lib/kubelet/pods/ad17376e-106a-4aed-abd5-bbc9a9737b11/volumes" Dec 10 10:03:22 crc kubenswrapper[4715]: I1210 10:03:22.593153 4715 scope.go:117] "RemoveContainer" containerID="1a48560f27a8b6743404f92c71729c3d1067c20585ceae502096c59d41300cb0" Dec 10 10:03:22 crc kubenswrapper[4715]: I1210 10:03:22.643345 4715 scope.go:117] "RemoveContainer" containerID="c51c3b816413cb9c1bd195a8e5ae598869eacdc3310e4b370b5d49c164de15ad" Dec 10 10:03:22 crc kubenswrapper[4715]: I1210 10:03:22.691291 4715 scope.go:117] "RemoveContainer" containerID="9ef9793a46d67594228c5ad70a237a786149e60b2da549a0aad6cb2b5bcc7930" Dec 10 10:03:22 crc kubenswrapper[4715]: I1210 10:03:22.749878 4715 scope.go:117] "RemoveContainer" containerID="069f829a7a1f67a425ce726262e67f467cf00ece44d7379725c41a6b91c5c453" Dec 10 10:03:25 crc kubenswrapper[4715]: I1210 10:03:25.526733 4715 generic.go:334] "Generic (PLEG): container finished" podID="42754cce-1e7d-416d-a1fa-118ae307880f" containerID="57fb13097d7d9ff6745a4324207689c41beea2e176610c911ef1bcf340cf3c24" exitCode=0 Dec 10 10:03:25 crc kubenswrapper[4715]: I1210 10:03:25.526856 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" event={"ID":"42754cce-1e7d-416d-a1fa-118ae307880f","Type":"ContainerDied","Data":"57fb13097d7d9ff6745a4324207689c41beea2e176610c911ef1bcf340cf3c24"} Dec 10 10:03:25 crc kubenswrapper[4715]: I1210 10:03:25.604424 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:03:25 crc kubenswrapper[4715]: E1210 10:03:25.604744 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:03:26 crc kubenswrapper[4715]: I1210 10:03:26.958881 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.054370 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-ssh-key\") pod \"42754cce-1e7d-416d-a1fa-118ae307880f\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.054550 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-inventory\") pod \"42754cce-1e7d-416d-a1fa-118ae307880f\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.054613 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh6kr\" (UniqueName: \"kubernetes.io/projected/42754cce-1e7d-416d-a1fa-118ae307880f-kube-api-access-xh6kr\") pod \"42754cce-1e7d-416d-a1fa-118ae307880f\" (UID: \"42754cce-1e7d-416d-a1fa-118ae307880f\") " Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.060436 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42754cce-1e7d-416d-a1fa-118ae307880f-kube-api-access-xh6kr" (OuterVolumeSpecName: "kube-api-access-xh6kr") pod "42754cce-1e7d-416d-a1fa-118ae307880f" (UID: "42754cce-1e7d-416d-a1fa-118ae307880f"). InnerVolumeSpecName "kube-api-access-xh6kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.082091 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-inventory" (OuterVolumeSpecName: "inventory") pod "42754cce-1e7d-416d-a1fa-118ae307880f" (UID: "42754cce-1e7d-416d-a1fa-118ae307880f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.082826 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "42754cce-1e7d-416d-a1fa-118ae307880f" (UID: "42754cce-1e7d-416d-a1fa-118ae307880f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.156506 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.156549 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42754cce-1e7d-416d-a1fa-118ae307880f-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.156563 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh6kr\" (UniqueName: \"kubernetes.io/projected/42754cce-1e7d-416d-a1fa-118ae307880f-kube-api-access-xh6kr\") on node \"crc\" DevicePath \"\"" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.550660 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" event={"ID":"42754cce-1e7d-416d-a1fa-118ae307880f","Type":"ContainerDied","Data":"f44e5f8ec97b4a01e5f3731d3769385813b005383bcf6db6a14f8abad2e95c42"} Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.550715 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f44e5f8ec97b4a01e5f3731d3769385813b005383bcf6db6a14f8abad2e95c42" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.550720 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.644180 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m"] Dec 10 10:03:27 crc kubenswrapper[4715]: E1210 10:03:27.644974 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerName="extract-utilities" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.644992 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerName="extract-utilities" Dec 10 10:03:27 crc kubenswrapper[4715]: E1210 10:03:27.645013 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42754cce-1e7d-416d-a1fa-118ae307880f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.645047 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="42754cce-1e7d-416d-a1fa-118ae307880f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 10:03:27 crc kubenswrapper[4715]: E1210 10:03:27.645061 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerName="extract-content" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.645067 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerName="extract-content" Dec 10 10:03:27 crc kubenswrapper[4715]: E1210 10:03:27.645080 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerName="registry-server" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.645085 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerName="registry-server" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.645292 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad17376e-106a-4aed-abd5-bbc9a9737b11" containerName="registry-server" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.645305 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="42754cce-1e7d-416d-a1fa-118ae307880f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.646001 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.650274 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.650447 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.650595 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.652586 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.658460 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m"] Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.674270 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.674470 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwdcp\" (UniqueName: \"kubernetes.io/projected/dc964814-44b2-4f7c-b4f2-7a9962758553-kube-api-access-fwdcp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.674501 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.775978 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.776207 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwdcp\" (UniqueName: \"kubernetes.io/projected/dc964814-44b2-4f7c-b4f2-7a9962758553-kube-api-access-fwdcp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.776241 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.781185 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.783268 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.802818 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwdcp\" (UniqueName: \"kubernetes.io/projected/dc964814-44b2-4f7c-b4f2-7a9962758553-kube-api-access-fwdcp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:27 crc kubenswrapper[4715]: I1210 10:03:27.973573 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:28 crc kubenswrapper[4715]: I1210 10:03:28.500187 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m"] Dec 10 10:03:28 crc kubenswrapper[4715]: W1210 10:03:28.504022 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc964814_44b2_4f7c_b4f2_7a9962758553.slice/crio-acf6248fdc4d2cb59b5e89e24650d659e5554f9a5e44bd7bcd6d73b5a5160218 WatchSource:0}: Error finding container acf6248fdc4d2cb59b5e89e24650d659e5554f9a5e44bd7bcd6d73b5a5160218: Status 404 returned error can't find the container with id acf6248fdc4d2cb59b5e89e24650d659e5554f9a5e44bd7bcd6d73b5a5160218 Dec 10 10:03:28 crc kubenswrapper[4715]: I1210 10:03:28.559232 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" event={"ID":"dc964814-44b2-4f7c-b4f2-7a9962758553","Type":"ContainerStarted","Data":"acf6248fdc4d2cb59b5e89e24650d659e5554f9a5e44bd7bcd6d73b5a5160218"} Dec 10 10:03:29 crc kubenswrapper[4715]: I1210 10:03:29.569658 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" event={"ID":"dc964814-44b2-4f7c-b4f2-7a9962758553","Type":"ContainerStarted","Data":"6a347362498e9e9d200a3f1ff54d103878572ec741867af1540d5f3f44641338"} Dec 10 10:03:29 crc kubenswrapper[4715]: I1210 10:03:29.590947 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" podStartSLOduration=2.082711625 podStartE2EDuration="2.590930476s" podCreationTimestamp="2025-12-10 10:03:27 +0000 UTC" firstStartedPulling="2025-12-10 10:03:28.506053424 +0000 UTC m=+1771.249599675" lastFinishedPulling="2025-12-10 10:03:29.014272275 +0000 UTC m=+1771.757818526" observedRunningTime="2025-12-10 10:03:29.581480907 +0000 UTC m=+1772.325027178" watchObservedRunningTime="2025-12-10 10:03:29.590930476 +0000 UTC m=+1772.334476727" Dec 10 10:03:34 crc kubenswrapper[4715]: I1210 10:03:34.633139 4715 generic.go:334] "Generic (PLEG): container finished" podID="dc964814-44b2-4f7c-b4f2-7a9962758553" containerID="6a347362498e9e9d200a3f1ff54d103878572ec741867af1540d5f3f44641338" exitCode=0 Dec 10 10:03:34 crc kubenswrapper[4715]: I1210 10:03:34.633373 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" event={"ID":"dc964814-44b2-4f7c-b4f2-7a9962758553","Type":"ContainerDied","Data":"6a347362498e9e9d200a3f1ff54d103878572ec741867af1540d5f3f44641338"} Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.066535 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.238175 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-ssh-key\") pod \"dc964814-44b2-4f7c-b4f2-7a9962758553\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.238302 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwdcp\" (UniqueName: \"kubernetes.io/projected/dc964814-44b2-4f7c-b4f2-7a9962758553-kube-api-access-fwdcp\") pod \"dc964814-44b2-4f7c-b4f2-7a9962758553\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.238376 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-inventory\") pod \"dc964814-44b2-4f7c-b4f2-7a9962758553\" (UID: \"dc964814-44b2-4f7c-b4f2-7a9962758553\") " Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.252780 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc964814-44b2-4f7c-b4f2-7a9962758553-kube-api-access-fwdcp" (OuterVolumeSpecName: "kube-api-access-fwdcp") pod "dc964814-44b2-4f7c-b4f2-7a9962758553" (UID: "dc964814-44b2-4f7c-b4f2-7a9962758553"). InnerVolumeSpecName "kube-api-access-fwdcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.269418 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-inventory" (OuterVolumeSpecName: "inventory") pod "dc964814-44b2-4f7c-b4f2-7a9962758553" (UID: "dc964814-44b2-4f7c-b4f2-7a9962758553"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.269993 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dc964814-44b2-4f7c-b4f2-7a9962758553" (UID: "dc964814-44b2-4f7c-b4f2-7a9962758553"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.341510 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.341547 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwdcp\" (UniqueName: \"kubernetes.io/projected/dc964814-44b2-4f7c-b4f2-7a9962758553-kube-api-access-fwdcp\") on node \"crc\" DevicePath \"\"" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.341562 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc964814-44b2-4f7c-b4f2-7a9962758553-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.654111 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" event={"ID":"dc964814-44b2-4f7c-b4f2-7a9962758553","Type":"ContainerDied","Data":"acf6248fdc4d2cb59b5e89e24650d659e5554f9a5e44bd7bcd6d73b5a5160218"} Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.654679 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acf6248fdc4d2cb59b5e89e24650d659e5554f9a5e44bd7bcd6d73b5a5160218" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.654205 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.781715 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4"] Dec 10 10:03:36 crc kubenswrapper[4715]: E1210 10:03:36.782226 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc964814-44b2-4f7c-b4f2-7a9962758553" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.782245 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc964814-44b2-4f7c-b4f2-7a9962758553" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.782453 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc964814-44b2-4f7c-b4f2-7a9962758553" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.783136 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.784986 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.784997 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.785433 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.785560 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.792021 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4"] Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.956260 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-h82b4\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.956342 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trs5l\" (UniqueName: \"kubernetes.io/projected/9692f891-f3d4-412b-92bd-d154842e0d79-kube-api-access-trs5l\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-h82b4\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:36 crc kubenswrapper[4715]: I1210 10:03:36.956440 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-h82b4\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:37 crc kubenswrapper[4715]: I1210 10:03:37.058608 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-h82b4\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:37 crc kubenswrapper[4715]: I1210 10:03:37.058715 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trs5l\" (UniqueName: \"kubernetes.io/projected/9692f891-f3d4-412b-92bd-d154842e0d79-kube-api-access-trs5l\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-h82b4\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:37 crc kubenswrapper[4715]: I1210 10:03:37.058849 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-h82b4\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:37 crc kubenswrapper[4715]: I1210 10:03:37.062790 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-h82b4\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:37 crc kubenswrapper[4715]: I1210 10:03:37.068832 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-h82b4\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:37 crc kubenswrapper[4715]: I1210 10:03:37.077610 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trs5l\" (UniqueName: \"kubernetes.io/projected/9692f891-f3d4-412b-92bd-d154842e0d79-kube-api-access-trs5l\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-h82b4\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:37 crc kubenswrapper[4715]: I1210 10:03:37.099838 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:03:37 crc kubenswrapper[4715]: I1210 10:03:37.601162 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4"] Dec 10 10:03:37 crc kubenswrapper[4715]: I1210 10:03:37.665899 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" event={"ID":"9692f891-f3d4-412b-92bd-d154842e0d79","Type":"ContainerStarted","Data":"9b3ec0b285c6d3a01f081262851eaa479e1907c667abbf79a0c90be0f91e2f21"} Dec 10 10:03:38 crc kubenswrapper[4715]: I1210 10:03:38.605405 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:03:38 crc kubenswrapper[4715]: E1210 10:03:38.607577 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:03:38 crc kubenswrapper[4715]: I1210 10:03:38.676327 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" event={"ID":"9692f891-f3d4-412b-92bd-d154842e0d79","Type":"ContainerStarted","Data":"1a25c793041269d27c3db01da9d010fbcca2907119fa4fddc07ded53236ecde7"} Dec 10 10:03:38 crc kubenswrapper[4715]: I1210 10:03:38.720601 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" podStartSLOduration=2.242738585 podStartE2EDuration="2.720572891s" podCreationTimestamp="2025-12-10 10:03:36 +0000 UTC" firstStartedPulling="2025-12-10 10:03:37.604937101 +0000 UTC m=+1780.348483352" lastFinishedPulling="2025-12-10 10:03:38.082771387 +0000 UTC m=+1780.826317658" observedRunningTime="2025-12-10 10:03:38.69454454 +0000 UTC m=+1781.438090791" watchObservedRunningTime="2025-12-10 10:03:38.720572891 +0000 UTC m=+1781.464119152" Dec 10 10:03:45 crc kubenswrapper[4715]: I1210 10:03:45.042682 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-s6r2g"] Dec 10 10:03:45 crc kubenswrapper[4715]: I1210 10:03:45.058560 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-s6r2g"] Dec 10 10:03:45 crc kubenswrapper[4715]: I1210 10:03:45.617018 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90981cd9-8647-4ee0-abc4-fdf840af114f" path="/var/lib/kubelet/pods/90981cd9-8647-4ee0-abc4-fdf840af114f/volumes" Dec 10 10:03:46 crc kubenswrapper[4715]: I1210 10:03:46.037091 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-ccce-account-create-update-8pckm"] Dec 10 10:03:46 crc kubenswrapper[4715]: I1210 10:03:46.044870 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-l8nbx"] Dec 10 10:03:46 crc kubenswrapper[4715]: I1210 10:03:46.053163 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-ccce-account-create-update-8pckm"] Dec 10 10:03:46 crc kubenswrapper[4715]: I1210 10:03:46.062338 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-dvpzw"] Dec 10 10:03:46 crc kubenswrapper[4715]: I1210 10:03:46.070232 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-b909-account-create-update-f7lvf"] Dec 10 10:03:46 crc kubenswrapper[4715]: I1210 10:03:46.078797 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-1365-account-create-update-h5zsg"] Dec 10 10:03:46 crc kubenswrapper[4715]: I1210 10:03:46.087111 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-l8nbx"] Dec 10 10:03:46 crc kubenswrapper[4715]: I1210 10:03:46.095077 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-b909-account-create-update-f7lvf"] Dec 10 10:03:46 crc kubenswrapper[4715]: I1210 10:03:46.104222 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-dvpzw"] Dec 10 10:03:46 crc kubenswrapper[4715]: I1210 10:03:46.114139 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-1365-account-create-update-h5zsg"] Dec 10 10:03:47 crc kubenswrapper[4715]: I1210 10:03:47.615641 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="284a9082-8b05-470c-93f3-1623b2a569dc" path="/var/lib/kubelet/pods/284a9082-8b05-470c-93f3-1623b2a569dc/volumes" Dec 10 10:03:47 crc kubenswrapper[4715]: I1210 10:03:47.616294 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a223f35-4f28-4207-b900-4738035524c1" path="/var/lib/kubelet/pods/8a223f35-4f28-4207-b900-4738035524c1/volumes" Dec 10 10:03:47 crc kubenswrapper[4715]: I1210 10:03:47.616933 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5bc47a9-7f16-4f49-b2e1-904dcdae3b52" path="/var/lib/kubelet/pods/b5bc47a9-7f16-4f49-b2e1-904dcdae3b52/volumes" Dec 10 10:03:47 crc kubenswrapper[4715]: I1210 10:03:47.617461 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b6c95a-d6f0-4753-bcaa-08cecbe2289f" path="/var/lib/kubelet/pods/d7b6c95a-d6f0-4753-bcaa-08cecbe2289f/volumes" Dec 10 10:03:47 crc kubenswrapper[4715]: I1210 10:03:47.618545 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e97c28af-aa74-4d5e-9c58-6bf64e02c4df" path="/var/lib/kubelet/pods/e97c28af-aa74-4d5e-9c58-6bf64e02c4df/volumes" Dec 10 10:03:52 crc kubenswrapper[4715]: I1210 10:03:52.604975 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:03:52 crc kubenswrapper[4715]: E1210 10:03:52.605889 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:04:05 crc kubenswrapper[4715]: I1210 10:04:05.605350 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:04:05 crc kubenswrapper[4715]: E1210 10:04:05.606024 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:04:16 crc kubenswrapper[4715]: I1210 10:04:16.026534 4715 generic.go:334] "Generic (PLEG): container finished" podID="9692f891-f3d4-412b-92bd-d154842e0d79" containerID="1a25c793041269d27c3db01da9d010fbcca2907119fa4fddc07ded53236ecde7" exitCode=0 Dec 10 10:04:16 crc kubenswrapper[4715]: I1210 10:04:16.026630 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" event={"ID":"9692f891-f3d4-412b-92bd-d154842e0d79","Type":"ContainerDied","Data":"1a25c793041269d27c3db01da9d010fbcca2907119fa4fddc07ded53236ecde7"} Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.055519 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kkzgm"] Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.070288 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kkzgm"] Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.480417 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.597122 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-inventory\") pod \"9692f891-f3d4-412b-92bd-d154842e0d79\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.597341 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-ssh-key\") pod \"9692f891-f3d4-412b-92bd-d154842e0d79\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.597372 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trs5l\" (UniqueName: \"kubernetes.io/projected/9692f891-f3d4-412b-92bd-d154842e0d79-kube-api-access-trs5l\") pod \"9692f891-f3d4-412b-92bd-d154842e0d79\" (UID: \"9692f891-f3d4-412b-92bd-d154842e0d79\") " Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.604493 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9692f891-f3d4-412b-92bd-d154842e0d79-kube-api-access-trs5l" (OuterVolumeSpecName: "kube-api-access-trs5l") pod "9692f891-f3d4-412b-92bd-d154842e0d79" (UID: "9692f891-f3d4-412b-92bd-d154842e0d79"). InnerVolumeSpecName "kube-api-access-trs5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.620014 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c7977c2-7055-41e5-8d3a-ad1e652186bc" path="/var/lib/kubelet/pods/8c7977c2-7055-41e5-8d3a-ad1e652186bc/volumes" Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.627478 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9692f891-f3d4-412b-92bd-d154842e0d79" (UID: "9692f891-f3d4-412b-92bd-d154842e0d79"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.627858 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-inventory" (OuterVolumeSpecName: "inventory") pod "9692f891-f3d4-412b-92bd-d154842e0d79" (UID: "9692f891-f3d4-412b-92bd-d154842e0d79"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.699566 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.699877 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9692f891-f3d4-412b-92bd-d154842e0d79-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:04:17 crc kubenswrapper[4715]: I1210 10:04:17.699887 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trs5l\" (UniqueName: \"kubernetes.io/projected/9692f891-f3d4-412b-92bd-d154842e0d79-kube-api-access-trs5l\") on node \"crc\" DevicePath \"\"" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.047702 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" event={"ID":"9692f891-f3d4-412b-92bd-d154842e0d79","Type":"ContainerDied","Data":"9b3ec0b285c6d3a01f081262851eaa479e1907c667abbf79a0c90be0f91e2f21"} Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.047744 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b3ec0b285c6d3a01f081262851eaa479e1907c667abbf79a0c90be0f91e2f21" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.047764 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-h82b4" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.158859 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96"] Dec 10 10:04:18 crc kubenswrapper[4715]: E1210 10:04:18.160051 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9692f891-f3d4-412b-92bd-d154842e0d79" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.160100 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9692f891-f3d4-412b-92bd-d154842e0d79" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.160588 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9692f891-f3d4-412b-92bd-d154842e0d79" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.161608 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.165130 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.165533 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.166653 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.166902 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.188260 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96"] Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.310598 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rxk96\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.310701 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhnbz\" (UniqueName: \"kubernetes.io/projected/c952af8b-0155-480f-a60f-399b9ec919c8-kube-api-access-xhnbz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rxk96\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.310781 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rxk96\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.412522 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rxk96\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.412618 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhnbz\" (UniqueName: \"kubernetes.io/projected/c952af8b-0155-480f-a60f-399b9ec919c8-kube-api-access-xhnbz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rxk96\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.412689 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rxk96\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.417788 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rxk96\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.425984 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rxk96\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.429223 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhnbz\" (UniqueName: \"kubernetes.io/projected/c952af8b-0155-480f-a60f-399b9ec919c8-kube-api-access-xhnbz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rxk96\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:18 crc kubenswrapper[4715]: I1210 10:04:18.487867 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:04:19 crc kubenswrapper[4715]: I1210 10:04:19.090143 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96"] Dec 10 10:04:20 crc kubenswrapper[4715]: I1210 10:04:20.066699 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" event={"ID":"c952af8b-0155-480f-a60f-399b9ec919c8","Type":"ContainerStarted","Data":"7fdf8781833ad64c80b1902673b716ac6ba41376048dd22f849bd15b4b8b0bfd"} Dec 10 10:04:20 crc kubenswrapper[4715]: I1210 10:04:20.067298 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" event={"ID":"c952af8b-0155-480f-a60f-399b9ec919c8","Type":"ContainerStarted","Data":"6a0019b5f3a97ec62a8216476c735c4b83687932a9fe969f3bc79853f85b8923"} Dec 10 10:04:20 crc kubenswrapper[4715]: I1210 10:04:20.093577 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" podStartSLOduration=1.653606158 podStartE2EDuration="2.093546763s" podCreationTimestamp="2025-12-10 10:04:18 +0000 UTC" firstStartedPulling="2025-12-10 10:04:19.100752045 +0000 UTC m=+1821.844298296" lastFinishedPulling="2025-12-10 10:04:19.54069265 +0000 UTC m=+1822.284238901" observedRunningTime="2025-12-10 10:04:20.083323642 +0000 UTC m=+1822.826869913" watchObservedRunningTime="2025-12-10 10:04:20.093546763 +0000 UTC m=+1822.837093054" Dec 10 10:04:20 crc kubenswrapper[4715]: I1210 10:04:20.604471 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:04:20 crc kubenswrapper[4715]: E1210 10:04:20.605048 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:04:22 crc kubenswrapper[4715]: I1210 10:04:22.877125 4715 scope.go:117] "RemoveContainer" containerID="61ff90633b7fde6006561463ae4b159e0ac49361f01f389e9161c22d5fd7982f" Dec 10 10:04:22 crc kubenswrapper[4715]: I1210 10:04:22.899288 4715 scope.go:117] "RemoveContainer" containerID="78038a3157086e42207fc0bf00fbc6e74570431fdca7963cacf975ac23572479" Dec 10 10:04:22 crc kubenswrapper[4715]: I1210 10:04:22.949941 4715 scope.go:117] "RemoveContainer" containerID="d35a3305f328c3e6e0e741d01efbfbd21dc4a0fda9f38e2536971940ee98434d" Dec 10 10:04:22 crc kubenswrapper[4715]: I1210 10:04:22.998947 4715 scope.go:117] "RemoveContainer" containerID="01e548c196e50c07af9e0cff406b4e663addfa3a7e025b7b064262d8ccac4cae" Dec 10 10:04:23 crc kubenswrapper[4715]: I1210 10:04:23.163207 4715 scope.go:117] "RemoveContainer" containerID="5714ec46d17f9b4828df90f75ed2d929ebaca9ce49cd494e8ab89ee2a5e49a88" Dec 10 10:04:23 crc kubenswrapper[4715]: I1210 10:04:23.183393 4715 scope.go:117] "RemoveContainer" containerID="a9df0962aca3eacbd049ad154eb47491a466190b45070a5f3c3b8d462739b884" Dec 10 10:04:23 crc kubenswrapper[4715]: I1210 10:04:23.243345 4715 scope.go:117] "RemoveContainer" containerID="078d33857fc129aa312a0999a6a5b26d216ea2d9d7f87e691900288c52c38f2a" Dec 10 10:04:35 crc kubenswrapper[4715]: I1210 10:04:35.605003 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:04:35 crc kubenswrapper[4715]: E1210 10:04:35.605778 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:04:43 crc kubenswrapper[4715]: I1210 10:04:43.047686 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-sbn8r"] Dec 10 10:04:43 crc kubenswrapper[4715]: I1210 10:04:43.066675 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-sbn8r"] Dec 10 10:04:43 crc kubenswrapper[4715]: I1210 10:04:43.625628 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9936c060-cb6f-4c92-895b-c0381d18085e" path="/var/lib/kubelet/pods/9936c060-cb6f-4c92-895b-c0381d18085e/volumes" Dec 10 10:04:46 crc kubenswrapper[4715]: I1210 10:04:46.025758 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-thxxv"] Dec 10 10:04:46 crc kubenswrapper[4715]: I1210 10:04:46.034771 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-thxxv"] Dec 10 10:04:46 crc kubenswrapper[4715]: I1210 10:04:46.605258 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:04:46 crc kubenswrapper[4715]: E1210 10:04:46.605503 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:04:47 crc kubenswrapper[4715]: I1210 10:04:47.627175 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af79009e-1855-4f22-8397-b1c762175b72" path="/var/lib/kubelet/pods/af79009e-1855-4f22-8397-b1c762175b72/volumes" Dec 10 10:04:59 crc kubenswrapper[4715]: I1210 10:04:59.605649 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:04:59 crc kubenswrapper[4715]: E1210 10:04:59.606526 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:05:12 crc kubenswrapper[4715]: I1210 10:05:12.615189 4715 generic.go:334] "Generic (PLEG): container finished" podID="c952af8b-0155-480f-a60f-399b9ec919c8" containerID="7fdf8781833ad64c80b1902673b716ac6ba41376048dd22f849bd15b4b8b0bfd" exitCode=0 Dec 10 10:05:12 crc kubenswrapper[4715]: I1210 10:05:12.615268 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" event={"ID":"c952af8b-0155-480f-a60f-399b9ec919c8","Type":"ContainerDied","Data":"7fdf8781833ad64c80b1902673b716ac6ba41376048dd22f849bd15b4b8b0bfd"} Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.025205 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.159425 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-inventory\") pod \"c952af8b-0155-480f-a60f-399b9ec919c8\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.159739 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhnbz\" (UniqueName: \"kubernetes.io/projected/c952af8b-0155-480f-a60f-399b9ec919c8-kube-api-access-xhnbz\") pod \"c952af8b-0155-480f-a60f-399b9ec919c8\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.160076 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-ssh-key\") pod \"c952af8b-0155-480f-a60f-399b9ec919c8\" (UID: \"c952af8b-0155-480f-a60f-399b9ec919c8\") " Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.165281 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c952af8b-0155-480f-a60f-399b9ec919c8-kube-api-access-xhnbz" (OuterVolumeSpecName: "kube-api-access-xhnbz") pod "c952af8b-0155-480f-a60f-399b9ec919c8" (UID: "c952af8b-0155-480f-a60f-399b9ec919c8"). InnerVolumeSpecName "kube-api-access-xhnbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.187701 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-inventory" (OuterVolumeSpecName: "inventory") pod "c952af8b-0155-480f-a60f-399b9ec919c8" (UID: "c952af8b-0155-480f-a60f-399b9ec919c8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.193807 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c952af8b-0155-480f-a60f-399b9ec919c8" (UID: "c952af8b-0155-480f-a60f-399b9ec919c8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.262572 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.262772 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c952af8b-0155-480f-a60f-399b9ec919c8-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.263183 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhnbz\" (UniqueName: \"kubernetes.io/projected/c952af8b-0155-480f-a60f-399b9ec919c8-kube-api-access-xhnbz\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.605113 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:05:14 crc kubenswrapper[4715]: E1210 10:05:14.605682 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.643046 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" event={"ID":"c952af8b-0155-480f-a60f-399b9ec919c8","Type":"ContainerDied","Data":"6a0019b5f3a97ec62a8216476c735c4b83687932a9fe969f3bc79853f85b8923"} Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.643093 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a0019b5f3a97ec62a8216476c735c4b83687932a9fe969f3bc79853f85b8923" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.643256 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rxk96" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.721932 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8h9tz"] Dec 10 10:05:14 crc kubenswrapper[4715]: E1210 10:05:14.722632 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c952af8b-0155-480f-a60f-399b9ec919c8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.722651 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c952af8b-0155-480f-a60f-399b9ec919c8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.722849 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c952af8b-0155-480f-a60f-399b9ec919c8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.723542 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.730411 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.730489 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.730604 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.730686 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.734440 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8h9tz"] Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.772051 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8h9tz\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.772113 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xbcw\" (UniqueName: \"kubernetes.io/projected/757f9102-c847-4fa9-b9a6-87495185cffe-kube-api-access-2xbcw\") pod \"ssh-known-hosts-edpm-deployment-8h9tz\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.772173 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8h9tz\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.873568 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8h9tz\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.873620 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xbcw\" (UniqueName: \"kubernetes.io/projected/757f9102-c847-4fa9-b9a6-87495185cffe-kube-api-access-2xbcw\") pod \"ssh-known-hosts-edpm-deployment-8h9tz\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.873699 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8h9tz\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.878387 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8h9tz\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.879985 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8h9tz\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:14 crc kubenswrapper[4715]: I1210 10:05:14.890614 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xbcw\" (UniqueName: \"kubernetes.io/projected/757f9102-c847-4fa9-b9a6-87495185cffe-kube-api-access-2xbcw\") pod \"ssh-known-hosts-edpm-deployment-8h9tz\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:15 crc kubenswrapper[4715]: I1210 10:05:15.046563 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:15 crc kubenswrapper[4715]: I1210 10:05:15.575701 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8h9tz"] Dec 10 10:05:15 crc kubenswrapper[4715]: W1210 10:05:15.580467 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod757f9102_c847_4fa9_b9a6_87495185cffe.slice/crio-9e1d552f2b73ff5ac75afd2cf7e3c98af47c15616726d1e74c35e6f226afa9c7 WatchSource:0}: Error finding container 9e1d552f2b73ff5ac75afd2cf7e3c98af47c15616726d1e74c35e6f226afa9c7: Status 404 returned error can't find the container with id 9e1d552f2b73ff5ac75afd2cf7e3c98af47c15616726d1e74c35e6f226afa9c7 Dec 10 10:05:15 crc kubenswrapper[4715]: I1210 10:05:15.583411 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 10:05:15 crc kubenswrapper[4715]: I1210 10:05:15.652571 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" event={"ID":"757f9102-c847-4fa9-b9a6-87495185cffe","Type":"ContainerStarted","Data":"9e1d552f2b73ff5ac75afd2cf7e3c98af47c15616726d1e74c35e6f226afa9c7"} Dec 10 10:05:16 crc kubenswrapper[4715]: I1210 10:05:16.662300 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" event={"ID":"757f9102-c847-4fa9-b9a6-87495185cffe","Type":"ContainerStarted","Data":"e49d529401f0ea8077a1fb0e29bfc62f60db902f063d711569e9bb12bc543fe8"} Dec 10 10:05:16 crc kubenswrapper[4715]: I1210 10:05:16.702361 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" podStartSLOduration=2.174196604 podStartE2EDuration="2.702338213s" podCreationTimestamp="2025-12-10 10:05:14 +0000 UTC" firstStartedPulling="2025-12-10 10:05:15.583194274 +0000 UTC m=+1878.326740525" lastFinishedPulling="2025-12-10 10:05:16.111335883 +0000 UTC m=+1878.854882134" observedRunningTime="2025-12-10 10:05:16.683191447 +0000 UTC m=+1879.426737698" watchObservedRunningTime="2025-12-10 10:05:16.702338213 +0000 UTC m=+1879.445884464" Dec 10 10:05:23 crc kubenswrapper[4715]: I1210 10:05:23.395242 4715 scope.go:117] "RemoveContainer" containerID="2dc1de321a39040ba4cb0ada4179d63413bbe3f996f2449d85212d2663e1099d" Dec 10 10:05:23 crc kubenswrapper[4715]: I1210 10:05:23.461215 4715 scope.go:117] "RemoveContainer" containerID="6d9766176de8ce5757d3983f0f9746dce51ce47c6ecc7e1151cf20b35bc7c17a" Dec 10 10:05:23 crc kubenswrapper[4715]: I1210 10:05:23.735613 4715 generic.go:334] "Generic (PLEG): container finished" podID="757f9102-c847-4fa9-b9a6-87495185cffe" containerID="e49d529401f0ea8077a1fb0e29bfc62f60db902f063d711569e9bb12bc543fe8" exitCode=0 Dec 10 10:05:23 crc kubenswrapper[4715]: I1210 10:05:23.735657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" event={"ID":"757f9102-c847-4fa9-b9a6-87495185cffe","Type":"ContainerDied","Data":"e49d529401f0ea8077a1fb0e29bfc62f60db902f063d711569e9bb12bc543fe8"} Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.214454 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.327125 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-ssh-key-openstack-edpm-ipam\") pod \"757f9102-c847-4fa9-b9a6-87495185cffe\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.327492 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xbcw\" (UniqueName: \"kubernetes.io/projected/757f9102-c847-4fa9-b9a6-87495185cffe-kube-api-access-2xbcw\") pod \"757f9102-c847-4fa9-b9a6-87495185cffe\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.327533 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-inventory-0\") pod \"757f9102-c847-4fa9-b9a6-87495185cffe\" (UID: \"757f9102-c847-4fa9-b9a6-87495185cffe\") " Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.333482 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/757f9102-c847-4fa9-b9a6-87495185cffe-kube-api-access-2xbcw" (OuterVolumeSpecName: "kube-api-access-2xbcw") pod "757f9102-c847-4fa9-b9a6-87495185cffe" (UID: "757f9102-c847-4fa9-b9a6-87495185cffe"). InnerVolumeSpecName "kube-api-access-2xbcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.354704 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "757f9102-c847-4fa9-b9a6-87495185cffe" (UID: "757f9102-c847-4fa9-b9a6-87495185cffe"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.356234 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "757f9102-c847-4fa9-b9a6-87495185cffe" (UID: "757f9102-c847-4fa9-b9a6-87495185cffe"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.430128 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xbcw\" (UniqueName: \"kubernetes.io/projected/757f9102-c847-4fa9-b9a6-87495185cffe-kube-api-access-2xbcw\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.430204 4715 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.430222 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/757f9102-c847-4fa9-b9a6-87495185cffe-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.754303 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" event={"ID":"757f9102-c847-4fa9-b9a6-87495185cffe","Type":"ContainerDied","Data":"9e1d552f2b73ff5ac75afd2cf7e3c98af47c15616726d1e74c35e6f226afa9c7"} Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.754365 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e1d552f2b73ff5ac75afd2cf7e3c98af47c15616726d1e74c35e6f226afa9c7" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.754423 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8h9tz" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.824252 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4"] Dec 10 10:05:25 crc kubenswrapper[4715]: E1210 10:05:25.825309 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="757f9102-c847-4fa9-b9a6-87495185cffe" containerName="ssh-known-hosts-edpm-deployment" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.825371 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="757f9102-c847-4fa9-b9a6-87495185cffe" containerName="ssh-known-hosts-edpm-deployment" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.825672 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="757f9102-c847-4fa9-b9a6-87495185cffe" containerName="ssh-known-hosts-edpm-deployment" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.826678 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.845168 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.845624 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.845653 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.845686 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.851567 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4"] Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.947857 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z2f8\" (UniqueName: \"kubernetes.io/projected/06a8246c-8107-4c74-b827-1c965bd840ec-kube-api-access-4z2f8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pz4s4\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.947950 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pz4s4\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:25 crc kubenswrapper[4715]: I1210 10:05:25.948126 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pz4s4\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:26 crc kubenswrapper[4715]: I1210 10:05:26.050407 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z2f8\" (UniqueName: \"kubernetes.io/projected/06a8246c-8107-4c74-b827-1c965bd840ec-kube-api-access-4z2f8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pz4s4\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:26 crc kubenswrapper[4715]: I1210 10:05:26.050479 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pz4s4\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:26 crc kubenswrapper[4715]: I1210 10:05:26.050539 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pz4s4\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:26 crc kubenswrapper[4715]: I1210 10:05:26.055680 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pz4s4\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:26 crc kubenswrapper[4715]: I1210 10:05:26.055692 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pz4s4\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:26 crc kubenswrapper[4715]: I1210 10:05:26.069273 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z2f8\" (UniqueName: \"kubernetes.io/projected/06a8246c-8107-4c74-b827-1c965bd840ec-kube-api-access-4z2f8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pz4s4\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:26 crc kubenswrapper[4715]: I1210 10:05:26.167261 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:26 crc kubenswrapper[4715]: I1210 10:05:26.676801 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4"] Dec 10 10:05:26 crc kubenswrapper[4715]: I1210 10:05:26.770234 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" event={"ID":"06a8246c-8107-4c74-b827-1c965bd840ec","Type":"ContainerStarted","Data":"73e80d0292e6b0719b4042f1540cd21fcf43a7ffd8ab0f47228c09f94a011352"} Dec 10 10:05:27 crc kubenswrapper[4715]: I1210 10:05:27.046526 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-xgcdc"] Dec 10 10:05:27 crc kubenswrapper[4715]: I1210 10:05:27.057805 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-xgcdc"] Dec 10 10:05:27 crc kubenswrapper[4715]: I1210 10:05:27.615842 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f07f968f-58f0-46c6-ac4e-d0cdcb966d10" path="/var/lib/kubelet/pods/f07f968f-58f0-46c6-ac4e-d0cdcb966d10/volumes" Dec 10 10:05:28 crc kubenswrapper[4715]: I1210 10:05:28.792369 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" event={"ID":"06a8246c-8107-4c74-b827-1c965bd840ec","Type":"ContainerStarted","Data":"54ae9197a30d2600b7d43741cdb58d6736fae0f38f8244bd47ff3f644c3bb12d"} Dec 10 10:05:28 crc kubenswrapper[4715]: I1210 10:05:28.817221 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" podStartSLOduration=2.613829902 podStartE2EDuration="3.81718396s" podCreationTimestamp="2025-12-10 10:05:25 +0000 UTC" firstStartedPulling="2025-12-10 10:05:26.670353467 +0000 UTC m=+1889.413899718" lastFinishedPulling="2025-12-10 10:05:27.873707525 +0000 UTC m=+1890.617253776" observedRunningTime="2025-12-10 10:05:28.809443209 +0000 UTC m=+1891.552989460" watchObservedRunningTime="2025-12-10 10:05:28.81718396 +0000 UTC m=+1891.560730201" Dec 10 10:05:29 crc kubenswrapper[4715]: I1210 10:05:29.605419 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:05:29 crc kubenswrapper[4715]: E1210 10:05:29.605944 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:05:36 crc kubenswrapper[4715]: I1210 10:05:36.868072 4715 generic.go:334] "Generic (PLEG): container finished" podID="06a8246c-8107-4c74-b827-1c965bd840ec" containerID="54ae9197a30d2600b7d43741cdb58d6736fae0f38f8244bd47ff3f644c3bb12d" exitCode=0 Dec 10 10:05:36 crc kubenswrapper[4715]: I1210 10:05:36.868164 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" event={"ID":"06a8246c-8107-4c74-b827-1c965bd840ec","Type":"ContainerDied","Data":"54ae9197a30d2600b7d43741cdb58d6736fae0f38f8244bd47ff3f644c3bb12d"} Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.325458 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.435596 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-ssh-key\") pod \"06a8246c-8107-4c74-b827-1c965bd840ec\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.435943 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z2f8\" (UniqueName: \"kubernetes.io/projected/06a8246c-8107-4c74-b827-1c965bd840ec-kube-api-access-4z2f8\") pod \"06a8246c-8107-4c74-b827-1c965bd840ec\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.436025 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-inventory\") pod \"06a8246c-8107-4c74-b827-1c965bd840ec\" (UID: \"06a8246c-8107-4c74-b827-1c965bd840ec\") " Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.441835 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06a8246c-8107-4c74-b827-1c965bd840ec-kube-api-access-4z2f8" (OuterVolumeSpecName: "kube-api-access-4z2f8") pod "06a8246c-8107-4c74-b827-1c965bd840ec" (UID: "06a8246c-8107-4c74-b827-1c965bd840ec"). InnerVolumeSpecName "kube-api-access-4z2f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.464767 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-inventory" (OuterVolumeSpecName: "inventory") pod "06a8246c-8107-4c74-b827-1c965bd840ec" (UID: "06a8246c-8107-4c74-b827-1c965bd840ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.465104 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "06a8246c-8107-4c74-b827-1c965bd840ec" (UID: "06a8246c-8107-4c74-b827-1c965bd840ec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.538261 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z2f8\" (UniqueName: \"kubernetes.io/projected/06a8246c-8107-4c74-b827-1c965bd840ec-kube-api-access-4z2f8\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.538297 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.538305 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06a8246c-8107-4c74-b827-1c965bd840ec-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.906635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" event={"ID":"06a8246c-8107-4c74-b827-1c965bd840ec","Type":"ContainerDied","Data":"73e80d0292e6b0719b4042f1540cd21fcf43a7ffd8ab0f47228c09f94a011352"} Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.906682 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73e80d0292e6b0719b4042f1540cd21fcf43a7ffd8ab0f47228c09f94a011352" Dec 10 10:05:38 crc kubenswrapper[4715]: I1210 10:05:38.906765 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pz4s4" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.011175 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547"] Dec 10 10:05:39 crc kubenswrapper[4715]: E1210 10:05:39.011753 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06a8246c-8107-4c74-b827-1c965bd840ec" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.011775 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="06a8246c-8107-4c74-b827-1c965bd840ec" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.012012 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="06a8246c-8107-4c74-b827-1c965bd840ec" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.013085 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.055435 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547"] Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.063988 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.064269 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.063988 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.064570 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.154892 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzkwj\" (UniqueName: \"kubernetes.io/projected/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-kube-api-access-tzkwj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sr547\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.154978 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sr547\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.155089 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sr547\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.256324 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzkwj\" (UniqueName: \"kubernetes.io/projected/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-kube-api-access-tzkwj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sr547\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.256410 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sr547\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.256481 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sr547\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.261190 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sr547\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.265393 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sr547\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.271768 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzkwj\" (UniqueName: \"kubernetes.io/projected/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-kube-api-access-tzkwj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sr547\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.387137 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.893495 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547"] Dec 10 10:05:39 crc kubenswrapper[4715]: I1210 10:05:39.921328 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" event={"ID":"6d1a7152-aa86-4400-b3bb-55ca096ef3f7","Type":"ContainerStarted","Data":"3a3f8abdca8a612f474064d375aaccd92a7f1916c95f853b580d8949404747e3"} Dec 10 10:05:40 crc kubenswrapper[4715]: I1210 10:05:40.605308 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:05:40 crc kubenswrapper[4715]: E1210 10:05:40.605947 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:05:40 crc kubenswrapper[4715]: I1210 10:05:40.936252 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" event={"ID":"6d1a7152-aa86-4400-b3bb-55ca096ef3f7","Type":"ContainerStarted","Data":"f1c7450a7ab3a8d0a2a2e5c0f40e3259c0d9e6e45eb73d40ccbce5c5657900c5"} Dec 10 10:05:40 crc kubenswrapper[4715]: I1210 10:05:40.957850 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" podStartSLOduration=2.4754423340000002 podStartE2EDuration="2.957823121s" podCreationTimestamp="2025-12-10 10:05:38 +0000 UTC" firstStartedPulling="2025-12-10 10:05:39.890431086 +0000 UTC m=+1902.633977337" lastFinishedPulling="2025-12-10 10:05:40.372811873 +0000 UTC m=+1903.116358124" observedRunningTime="2025-12-10 10:05:40.953400485 +0000 UTC m=+1903.696946776" watchObservedRunningTime="2025-12-10 10:05:40.957823121 +0000 UTC m=+1903.701369362" Dec 10 10:05:50 crc kubenswrapper[4715]: I1210 10:05:50.107549 4715 generic.go:334] "Generic (PLEG): container finished" podID="6d1a7152-aa86-4400-b3bb-55ca096ef3f7" containerID="f1c7450a7ab3a8d0a2a2e5c0f40e3259c0d9e6e45eb73d40ccbce5c5657900c5" exitCode=0 Dec 10 10:05:50 crc kubenswrapper[4715]: I1210 10:05:50.107638 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" event={"ID":"6d1a7152-aa86-4400-b3bb-55ca096ef3f7","Type":"ContainerDied","Data":"f1c7450a7ab3a8d0a2a2e5c0f40e3259c0d9e6e45eb73d40ccbce5c5657900c5"} Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.539044 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.612181 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:05:51 crc kubenswrapper[4715]: E1210 10:05:51.612430 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.711767 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-inventory\") pod \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.711856 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-ssh-key\") pod \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.711930 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzkwj\" (UniqueName: \"kubernetes.io/projected/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-kube-api-access-tzkwj\") pod \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\" (UID: \"6d1a7152-aa86-4400-b3bb-55ca096ef3f7\") " Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.718132 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-kube-api-access-tzkwj" (OuterVolumeSpecName: "kube-api-access-tzkwj") pod "6d1a7152-aa86-4400-b3bb-55ca096ef3f7" (UID: "6d1a7152-aa86-4400-b3bb-55ca096ef3f7"). InnerVolumeSpecName "kube-api-access-tzkwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.736878 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-inventory" (OuterVolumeSpecName: "inventory") pod "6d1a7152-aa86-4400-b3bb-55ca096ef3f7" (UID: "6d1a7152-aa86-4400-b3bb-55ca096ef3f7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.739271 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6d1a7152-aa86-4400-b3bb-55ca096ef3f7" (UID: "6d1a7152-aa86-4400-b3bb-55ca096ef3f7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.817197 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.817264 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzkwj\" (UniqueName: \"kubernetes.io/projected/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-kube-api-access-tzkwj\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:51 crc kubenswrapper[4715]: I1210 10:05:51.817280 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d1a7152-aa86-4400-b3bb-55ca096ef3f7-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.132651 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" event={"ID":"6d1a7152-aa86-4400-b3bb-55ca096ef3f7","Type":"ContainerDied","Data":"3a3f8abdca8a612f474064d375aaccd92a7f1916c95f853b580d8949404747e3"} Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.132684 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sr547" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.132699 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a3f8abdca8a612f474064d375aaccd92a7f1916c95f853b580d8949404747e3" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.240411 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk"] Dec 10 10:05:52 crc kubenswrapper[4715]: E1210 10:05:52.245573 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1a7152-aa86-4400-b3bb-55ca096ef3f7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.245609 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1a7152-aa86-4400-b3bb-55ca096ef3f7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.247749 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1a7152-aa86-4400-b3bb-55ca096ef3f7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.250555 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.369435 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.370479 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.374382 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.374641 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.375875 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.376089 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.376242 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.376846 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.408293 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk"] Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470181 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470245 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470275 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470311 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470346 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470374 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470519 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470557 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470585 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470714 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470758 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470795 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470836 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4dtz\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-kube-api-access-f4dtz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.470906 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572295 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572368 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572390 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572415 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572439 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572465 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572486 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572512 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572527 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572554 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572612 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572654 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572685 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.572719 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4dtz\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-kube-api-access-f4dtz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.576758 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.576786 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.578634 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.578631 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.578662 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.579107 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.579609 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.579776 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.581377 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.581417 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.582448 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.587925 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.591772 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.597355 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4dtz\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-kube-api-access-f4dtz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:52 crc kubenswrapper[4715]: I1210 10:05:52.677764 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:05:53 crc kubenswrapper[4715]: I1210 10:05:53.223628 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk"] Dec 10 10:05:54 crc kubenswrapper[4715]: I1210 10:05:54.158018 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" event={"ID":"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9","Type":"ContainerStarted","Data":"4f76e32500b4700615ffecba049644294edfcda922822b084356bda18ab75b30"} Dec 10 10:05:54 crc kubenswrapper[4715]: I1210 10:05:54.158346 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" event={"ID":"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9","Type":"ContainerStarted","Data":"1359275ba6825470b15fa757040d8345491e554140aa9afb5e69625024629638"} Dec 10 10:05:54 crc kubenswrapper[4715]: I1210 10:05:54.186882 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" podStartSLOduration=1.717686697 podStartE2EDuration="2.186853286s" podCreationTimestamp="2025-12-10 10:05:52 +0000 UTC" firstStartedPulling="2025-12-10 10:05:53.23597239 +0000 UTC m=+1915.979518661" lastFinishedPulling="2025-12-10 10:05:53.705138999 +0000 UTC m=+1916.448685250" observedRunningTime="2025-12-10 10:05:54.180940237 +0000 UTC m=+1916.924486488" watchObservedRunningTime="2025-12-10 10:05:54.186853286 +0000 UTC m=+1916.930399537" Dec 10 10:06:06 crc kubenswrapper[4715]: I1210 10:06:06.605589 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:06:06 crc kubenswrapper[4715]: E1210 10:06:06.606404 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:06:19 crc kubenswrapper[4715]: I1210 10:06:19.605213 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:06:20 crc kubenswrapper[4715]: I1210 10:06:20.393802 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"5b2eee0605c28977d3538b338dcb2d691766610d0548ab30f1e772bdc0f6e9ba"} Dec 10 10:06:23 crc kubenswrapper[4715]: I1210 10:06:23.542749 4715 scope.go:117] "RemoveContainer" containerID="69f349fcb33c1d2a5a8883c637c243d58b1b07882c0e4513dd824035be60d472" Dec 10 10:06:32 crc kubenswrapper[4715]: I1210 10:06:32.513716 4715 generic.go:334] "Generic (PLEG): container finished" podID="0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" containerID="4f76e32500b4700615ffecba049644294edfcda922822b084356bda18ab75b30" exitCode=0 Dec 10 10:06:32 crc kubenswrapper[4715]: I1210 10:06:32.513798 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" event={"ID":"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9","Type":"ContainerDied","Data":"4f76e32500b4700615ffecba049644294edfcda922822b084356bda18ab75b30"} Dec 10 10:06:33 crc kubenswrapper[4715]: I1210 10:06:33.972489 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138308 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138395 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-neutron-metadata-combined-ca-bundle\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138421 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138458 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4dtz\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-kube-api-access-f4dtz\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138486 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ovn-combined-ca-bundle\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138538 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-repo-setup-combined-ca-bundle\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138584 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ssh-key\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138618 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-bootstrap-combined-ca-bundle\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138655 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-libvirt-combined-ca-bundle\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138693 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-telemetry-combined-ca-bundle\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138735 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138756 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-inventory\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138842 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.138871 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-nova-combined-ca-bundle\") pod \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\" (UID: \"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9\") " Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.144618 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.146379 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.146468 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-kube-api-access-f4dtz" (OuterVolumeSpecName: "kube-api-access-f4dtz") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "kube-api-access-f4dtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.147444 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.147474 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.147634 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.148800 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.148861 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.150026 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.150271 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.150783 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.151433 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.177097 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.178249 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-inventory" (OuterVolumeSpecName: "inventory") pod "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" (UID: "0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.242602 4715 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.242881 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243001 4715 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243074 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243154 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4dtz\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-kube-api-access-f4dtz\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243231 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243298 4715 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243376 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243480 4715 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243568 4715 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243639 4715 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243710 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243782 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.243857 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.532876 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" event={"ID":"0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9","Type":"ContainerDied","Data":"1359275ba6825470b15fa757040d8345491e554140aa9afb5e69625024629638"} Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.533192 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1359275ba6825470b15fa757040d8345491e554140aa9afb5e69625024629638" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.533142 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.727428 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq"] Dec 10 10:06:34 crc kubenswrapper[4715]: E1210 10:06:34.728219 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.728318 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.728690 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.729671 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.733162 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.733481 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.734810 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.735807 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.740007 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.767262 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq"] Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.857196 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lp42\" (UniqueName: \"kubernetes.io/projected/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-kube-api-access-4lp42\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.857347 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.857380 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.857406 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.857449 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.958662 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lp42\" (UniqueName: \"kubernetes.io/projected/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-kube-api-access-4lp42\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.958837 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.958871 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.958898 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.958967 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.959905 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.962975 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.962994 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.974973 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:34 crc kubenswrapper[4715]: I1210 10:06:34.975248 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lp42\" (UniqueName: \"kubernetes.io/projected/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-kube-api-access-4lp42\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-plxxq\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:35 crc kubenswrapper[4715]: I1210 10:06:35.065025 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:06:35 crc kubenswrapper[4715]: W1210 10:06:35.608408 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb70f0fbd_ec55_40e1_b61d_b34b1b92fa3b.slice/crio-5be7f04cc288f5849cf002aa562bcd829d82ecde7ab2b85b3d8588a70f5b6b3b WatchSource:0}: Error finding container 5be7f04cc288f5849cf002aa562bcd829d82ecde7ab2b85b3d8588a70f5b6b3b: Status 404 returned error can't find the container with id 5be7f04cc288f5849cf002aa562bcd829d82ecde7ab2b85b3d8588a70f5b6b3b Dec 10 10:06:35 crc kubenswrapper[4715]: I1210 10:06:35.730505 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq"] Dec 10 10:06:36 crc kubenswrapper[4715]: I1210 10:06:36.551172 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" event={"ID":"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b","Type":"ContainerStarted","Data":"0dee86c15ccf1f5f5707a6df6190dc5be9e50063a65634e82066b79a433d0a88"} Dec 10 10:06:36 crc kubenswrapper[4715]: I1210 10:06:36.551674 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" event={"ID":"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b","Type":"ContainerStarted","Data":"5be7f04cc288f5849cf002aa562bcd829d82ecde7ab2b85b3d8588a70f5b6b3b"} Dec 10 10:06:36 crc kubenswrapper[4715]: I1210 10:06:36.569674 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" podStartSLOduration=1.961712122 podStartE2EDuration="2.569655097s" podCreationTimestamp="2025-12-10 10:06:34 +0000 UTC" firstStartedPulling="2025-12-10 10:06:35.610725221 +0000 UTC m=+1958.354271472" lastFinishedPulling="2025-12-10 10:06:36.218668196 +0000 UTC m=+1958.962214447" observedRunningTime="2025-12-10 10:06:36.566101695 +0000 UTC m=+1959.309647956" watchObservedRunningTime="2025-12-10 10:06:36.569655097 +0000 UTC m=+1959.313201348" Dec 10 10:07:38 crc kubenswrapper[4715]: I1210 10:07:38.186580 4715 generic.go:334] "Generic (PLEG): container finished" podID="b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b" containerID="0dee86c15ccf1f5f5707a6df6190dc5be9e50063a65634e82066b79a433d0a88" exitCode=0 Dec 10 10:07:38 crc kubenswrapper[4715]: I1210 10:07:38.186683 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" event={"ID":"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b","Type":"ContainerDied","Data":"0dee86c15ccf1f5f5707a6df6190dc5be9e50063a65634e82066b79a433d0a88"} Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.655810 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.747722 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t2dq4"] Dec 10 10:07:39 crc kubenswrapper[4715]: E1210 10:07:39.748517 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.748546 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.748817 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.750817 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.772094 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2dq4"] Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.832346 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovn-combined-ca-bundle\") pod \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.832512 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-inventory\") pod \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.832555 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lp42\" (UniqueName: \"kubernetes.io/projected/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-kube-api-access-4lp42\") pod \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.832578 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovncontroller-config-0\") pod \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.832634 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ssh-key\") pod \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\" (UID: \"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b\") " Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.840500 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b" (UID: "b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.855228 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-kube-api-access-4lp42" (OuterVolumeSpecName: "kube-api-access-4lp42") pod "b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b" (UID: "b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b"). InnerVolumeSpecName "kube-api-access-4lp42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.865019 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b" (UID: "b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.872654 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-inventory" (OuterVolumeSpecName: "inventory") pod "b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b" (UID: "b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.883208 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b" (UID: "b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.934198 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtjmw\" (UniqueName: \"kubernetes.io/projected/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-kube-api-access-mtjmw\") pod \"certified-operators-t2dq4\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.934265 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-utilities\") pod \"certified-operators-t2dq4\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.934376 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-catalog-content\") pod \"certified-operators-t2dq4\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.934476 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.934492 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.934502 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lp42\" (UniqueName: \"kubernetes.io/projected/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-kube-api-access-4lp42\") on node \"crc\" DevicePath \"\"" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.934513 4715 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:07:39 crc kubenswrapper[4715]: I1210 10:07:39.934521 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.067944 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtjmw\" (UniqueName: \"kubernetes.io/projected/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-kube-api-access-mtjmw\") pod \"certified-operators-t2dq4\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.068071 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-utilities\") pod \"certified-operators-t2dq4\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.068199 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-catalog-content\") pod \"certified-operators-t2dq4\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.068700 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-utilities\") pod \"certified-operators-t2dq4\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.068871 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-catalog-content\") pod \"certified-operators-t2dq4\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.089383 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtjmw\" (UniqueName: \"kubernetes.io/projected/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-kube-api-access-mtjmw\") pod \"certified-operators-t2dq4\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.242453 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" event={"ID":"b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b","Type":"ContainerDied","Data":"5be7f04cc288f5849cf002aa562bcd829d82ecde7ab2b85b3d8588a70f5b6b3b"} Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.242496 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5be7f04cc288f5849cf002aa562bcd829d82ecde7ab2b85b3d8588a70f5b6b3b" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.242586 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-plxxq" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.326799 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6"] Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.328141 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.333434 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.334864 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.335817 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.335849 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.336396 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.336596 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.343788 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6"] Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.388554 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.389828 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.389889 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.390013 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.390064 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z86sh\" (UniqueName: \"kubernetes.io/projected/06251a0c-f042-4137-864b-c7003449b040-kube-api-access-z86sh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.390161 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.390192 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.496486 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.496543 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.496606 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.496653 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z86sh\" (UniqueName: \"kubernetes.io/projected/06251a0c-f042-4137-864b-c7003449b040-kube-api-access-z86sh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.496713 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.496750 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.503220 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.503550 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.504835 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.507725 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.509520 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.520780 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z86sh\" (UniqueName: \"kubernetes.io/projected/06251a0c-f042-4137-864b-c7003449b040-kube-api-access-z86sh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:40 crc kubenswrapper[4715]: I1210 10:07:40.697547 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:07:41 crc kubenswrapper[4715]: I1210 10:07:41.020487 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2dq4"] Dec 10 10:07:41 crc kubenswrapper[4715]: I1210 10:07:41.147840 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6"] Dec 10 10:07:41 crc kubenswrapper[4715]: I1210 10:07:41.256419 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" event={"ID":"06251a0c-f042-4137-864b-c7003449b040","Type":"ContainerStarted","Data":"457bbcf7d5e1234e3782cca857eb1f945501a07bd25e1e45268c0ccde2746604"} Dec 10 10:07:41 crc kubenswrapper[4715]: I1210 10:07:41.259337 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2dq4" event={"ID":"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74","Type":"ContainerStarted","Data":"a944b70231e171c0f3f9005c9f4d0437326ea387252a6417429f853cf0527f7c"} Dec 10 10:07:42 crc kubenswrapper[4715]: I1210 10:07:42.270049 4715 generic.go:334] "Generic (PLEG): container finished" podID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerID="0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4" exitCode=0 Dec 10 10:07:42 crc kubenswrapper[4715]: I1210 10:07:42.270166 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2dq4" event={"ID":"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74","Type":"ContainerDied","Data":"0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4"} Dec 10 10:07:42 crc kubenswrapper[4715]: I1210 10:07:42.272640 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" event={"ID":"06251a0c-f042-4137-864b-c7003449b040","Type":"ContainerStarted","Data":"cb335686582a687b477dfaf277bfe1c172e47075d497ccf61a1d5193d3f244e5"} Dec 10 10:07:42 crc kubenswrapper[4715]: I1210 10:07:42.310357 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" podStartSLOduration=1.620326143 podStartE2EDuration="2.310333763s" podCreationTimestamp="2025-12-10 10:07:40 +0000 UTC" firstStartedPulling="2025-12-10 10:07:41.163802705 +0000 UTC m=+2023.907348956" lastFinishedPulling="2025-12-10 10:07:41.853810325 +0000 UTC m=+2024.597356576" observedRunningTime="2025-12-10 10:07:42.305369391 +0000 UTC m=+2025.048915632" watchObservedRunningTime="2025-12-10 10:07:42.310333763 +0000 UTC m=+2025.053880014" Dec 10 10:07:43 crc kubenswrapper[4715]: I1210 10:07:43.301822 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2dq4" event={"ID":"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74","Type":"ContainerStarted","Data":"f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153"} Dec 10 10:07:44 crc kubenswrapper[4715]: I1210 10:07:44.313756 4715 generic.go:334] "Generic (PLEG): container finished" podID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerID="f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153" exitCode=0 Dec 10 10:07:44 crc kubenswrapper[4715]: I1210 10:07:44.313891 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2dq4" event={"ID":"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74","Type":"ContainerDied","Data":"f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153"} Dec 10 10:07:45 crc kubenswrapper[4715]: I1210 10:07:45.330233 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2dq4" event={"ID":"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74","Type":"ContainerStarted","Data":"1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085"} Dec 10 10:07:45 crc kubenswrapper[4715]: I1210 10:07:45.355698 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t2dq4" podStartSLOduration=3.846546938 podStartE2EDuration="6.355680507s" podCreationTimestamp="2025-12-10 10:07:39 +0000 UTC" firstStartedPulling="2025-12-10 10:07:42.271878774 +0000 UTC m=+2025.015425025" lastFinishedPulling="2025-12-10 10:07:44.781012323 +0000 UTC m=+2027.524558594" observedRunningTime="2025-12-10 10:07:45.347187035 +0000 UTC m=+2028.090733296" watchObservedRunningTime="2025-12-10 10:07:45.355680507 +0000 UTC m=+2028.099226758" Dec 10 10:07:50 crc kubenswrapper[4715]: I1210 10:07:50.389290 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:50 crc kubenswrapper[4715]: I1210 10:07:50.389947 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:50 crc kubenswrapper[4715]: I1210 10:07:50.472545 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:51 crc kubenswrapper[4715]: I1210 10:07:51.466821 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:52 crc kubenswrapper[4715]: I1210 10:07:52.727274 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2dq4"] Dec 10 10:07:53 crc kubenswrapper[4715]: I1210 10:07:53.401312 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t2dq4" podUID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerName="registry-server" containerID="cri-o://1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085" gracePeriod=2 Dec 10 10:07:53 crc kubenswrapper[4715]: I1210 10:07:53.958450 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.066163 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-catalog-content\") pod \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.066347 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtjmw\" (UniqueName: \"kubernetes.io/projected/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-kube-api-access-mtjmw\") pod \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.066586 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-utilities\") pod \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\" (UID: \"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74\") " Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.067531 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-utilities" (OuterVolumeSpecName: "utilities") pod "cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" (UID: "cfa73a7d-1c92-4a5b-896f-0bedb19ddc74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.071788 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-kube-api-access-mtjmw" (OuterVolumeSpecName: "kube-api-access-mtjmw") pod "cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" (UID: "cfa73a7d-1c92-4a5b-896f-0bedb19ddc74"). InnerVolumeSpecName "kube-api-access-mtjmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.113501 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" (UID: "cfa73a7d-1c92-4a5b-896f-0bedb19ddc74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.168552 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.168590 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtjmw\" (UniqueName: \"kubernetes.io/projected/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-kube-api-access-mtjmw\") on node \"crc\" DevicePath \"\"" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.168602 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.414109 4715 generic.go:334] "Generic (PLEG): container finished" podID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerID="1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085" exitCode=0 Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.414443 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2dq4" event={"ID":"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74","Type":"ContainerDied","Data":"1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085"} Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.414872 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2dq4" event={"ID":"cfa73a7d-1c92-4a5b-896f-0bedb19ddc74","Type":"ContainerDied","Data":"a944b70231e171c0f3f9005c9f4d0437326ea387252a6417429f853cf0527f7c"} Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.414898 4715 scope.go:117] "RemoveContainer" containerID="1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.414575 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2dq4" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.452550 4715 scope.go:117] "RemoveContainer" containerID="f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.462000 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2dq4"] Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.472280 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t2dq4"] Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.484373 4715 scope.go:117] "RemoveContainer" containerID="0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.517747 4715 scope.go:117] "RemoveContainer" containerID="1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085" Dec 10 10:07:54 crc kubenswrapper[4715]: E1210 10:07:54.518650 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085\": container with ID starting with 1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085 not found: ID does not exist" containerID="1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.518685 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085"} err="failed to get container status \"1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085\": rpc error: code = NotFound desc = could not find container \"1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085\": container with ID starting with 1ab52e70ba7da0b3ddd7dc6bc8752d1fb79d618e8c3471b1e046cfe8ddc24085 not found: ID does not exist" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.518704 4715 scope.go:117] "RemoveContainer" containerID="f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153" Dec 10 10:07:54 crc kubenswrapper[4715]: E1210 10:07:54.527291 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153\": container with ID starting with f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153 not found: ID does not exist" containerID="f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.527325 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153"} err="failed to get container status \"f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153\": rpc error: code = NotFound desc = could not find container \"f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153\": container with ID starting with f01381f87eb4f0fc4f41cfbe0f835750ba58952ae1920beb90fd0b68452a1153 not found: ID does not exist" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.527344 4715 scope.go:117] "RemoveContainer" containerID="0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4" Dec 10 10:07:54 crc kubenswrapper[4715]: E1210 10:07:54.527703 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4\": container with ID starting with 0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4 not found: ID does not exist" containerID="0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4" Dec 10 10:07:54 crc kubenswrapper[4715]: I1210 10:07:54.527726 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4"} err="failed to get container status \"0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4\": rpc error: code = NotFound desc = could not find container \"0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4\": container with ID starting with 0f92e6ad6917698ec94b5a240a4c199c4e3d47e0620ae601c93edbdedc245fd4 not found: ID does not exist" Dec 10 10:07:55 crc kubenswrapper[4715]: I1210 10:07:55.616392 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" path="/var/lib/kubelet/pods/cfa73a7d-1c92-4a5b-896f-0bedb19ddc74/volumes" Dec 10 10:08:30 crc kubenswrapper[4715]: I1210 10:08:30.785742 4715 generic.go:334] "Generic (PLEG): container finished" podID="06251a0c-f042-4137-864b-c7003449b040" containerID="cb335686582a687b477dfaf277bfe1c172e47075d497ccf61a1d5193d3f244e5" exitCode=0 Dec 10 10:08:30 crc kubenswrapper[4715]: I1210 10:08:30.785843 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" event={"ID":"06251a0c-f042-4137-864b-c7003449b040","Type":"ContainerDied","Data":"cb335686582a687b477dfaf277bfe1c172e47075d497ccf61a1d5193d3f244e5"} Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.230343 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.391117 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-nova-metadata-neutron-config-0\") pod \"06251a0c-f042-4137-864b-c7003449b040\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.391211 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-inventory\") pod \"06251a0c-f042-4137-864b-c7003449b040\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.391238 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-metadata-combined-ca-bundle\") pod \"06251a0c-f042-4137-864b-c7003449b040\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.391283 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-ssh-key\") pod \"06251a0c-f042-4137-864b-c7003449b040\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.391333 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z86sh\" (UniqueName: \"kubernetes.io/projected/06251a0c-f042-4137-864b-c7003449b040-kube-api-access-z86sh\") pod \"06251a0c-f042-4137-864b-c7003449b040\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.391420 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-ovn-metadata-agent-neutron-config-0\") pod \"06251a0c-f042-4137-864b-c7003449b040\" (UID: \"06251a0c-f042-4137-864b-c7003449b040\") " Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.397532 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06251a0c-f042-4137-864b-c7003449b040-kube-api-access-z86sh" (OuterVolumeSpecName: "kube-api-access-z86sh") pod "06251a0c-f042-4137-864b-c7003449b040" (UID: "06251a0c-f042-4137-864b-c7003449b040"). InnerVolumeSpecName "kube-api-access-z86sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.398425 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "06251a0c-f042-4137-864b-c7003449b040" (UID: "06251a0c-f042-4137-864b-c7003449b040"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.418538 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-inventory" (OuterVolumeSpecName: "inventory") pod "06251a0c-f042-4137-864b-c7003449b040" (UID: "06251a0c-f042-4137-864b-c7003449b040"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.420040 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "06251a0c-f042-4137-864b-c7003449b040" (UID: "06251a0c-f042-4137-864b-c7003449b040"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.420202 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "06251a0c-f042-4137-864b-c7003449b040" (UID: "06251a0c-f042-4137-864b-c7003449b040"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.427086 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "06251a0c-f042-4137-864b-c7003449b040" (UID: "06251a0c-f042-4137-864b-c7003449b040"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.494192 4715 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.494541 4715 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.494552 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.494562 4715 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.494572 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06251a0c-f042-4137-864b-c7003449b040-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.494580 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z86sh\" (UniqueName: \"kubernetes.io/projected/06251a0c-f042-4137-864b-c7003449b040-kube-api-access-z86sh\") on node \"crc\" DevicePath \"\"" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.804826 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" event={"ID":"06251a0c-f042-4137-864b-c7003449b040","Type":"ContainerDied","Data":"457bbcf7d5e1234e3782cca857eb1f945501a07bd25e1e45268c0ccde2746604"} Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.804863 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="457bbcf7d5e1234e3782cca857eb1f945501a07bd25e1e45268c0ccde2746604" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.804933 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.908449 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc"] Dec 10 10:08:32 crc kubenswrapper[4715]: E1210 10:08:32.909163 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06251a0c-f042-4137-864b-c7003449b040" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.909198 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="06251a0c-f042-4137-864b-c7003449b040" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 10:08:32 crc kubenswrapper[4715]: E1210 10:08:32.909225 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerName="extract-content" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.909238 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerName="extract-content" Dec 10 10:08:32 crc kubenswrapper[4715]: E1210 10:08:32.909282 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerName="extract-utilities" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.909293 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerName="extract-utilities" Dec 10 10:08:32 crc kubenswrapper[4715]: E1210 10:08:32.909340 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerName="registry-server" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.909351 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerName="registry-server" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.909666 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfa73a7d-1c92-4a5b-896f-0bedb19ddc74" containerName="registry-server" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.909744 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="06251a0c-f042-4137-864b-c7003449b040" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.910813 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.918489 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.918714 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.918857 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.918729 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.919052 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:08:32 crc kubenswrapper[4715]: I1210 10:08:32.922832 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc"] Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.116784 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sckcq\" (UniqueName: \"kubernetes.io/projected/cd86f221-0c52-4565-87ed-9d4c473aaaf4-kube-api-access-sckcq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.116939 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.116986 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.117019 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.117589 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.219959 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.220087 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.220151 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.220235 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.220319 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sckcq\" (UniqueName: \"kubernetes.io/projected/cd86f221-0c52-4565-87ed-9d4c473aaaf4-kube-api-access-sckcq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.224523 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.225019 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.225657 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.225859 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.242015 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sckcq\" (UniqueName: \"kubernetes.io/projected/cd86f221-0c52-4565-87ed-9d4c473aaaf4-kube-api-access-sckcq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-959dc\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:33 crc kubenswrapper[4715]: I1210 10:08:33.541721 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:08:34 crc kubenswrapper[4715]: I1210 10:08:34.185219 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc"] Dec 10 10:08:34 crc kubenswrapper[4715]: I1210 10:08:34.827848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" event={"ID":"cd86f221-0c52-4565-87ed-9d4c473aaaf4","Type":"ContainerStarted","Data":"733041d536eabd55811ac98a8a1a3fea750f11568fa6cedc996d7137591b6f06"} Dec 10 10:08:35 crc kubenswrapper[4715]: I1210 10:08:35.839798 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" event={"ID":"cd86f221-0c52-4565-87ed-9d4c473aaaf4","Type":"ContainerStarted","Data":"6e7a58fc5a75316e2ef9ddc59edc4e5309cb863dbae332b492d19af73fd45b73"} Dec 10 10:08:35 crc kubenswrapper[4715]: I1210 10:08:35.868659 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" podStartSLOduration=2.872849293 podStartE2EDuration="3.868634473s" podCreationTimestamp="2025-12-10 10:08:32 +0000 UTC" firstStartedPulling="2025-12-10 10:08:34.193339473 +0000 UTC m=+2076.936885724" lastFinishedPulling="2025-12-10 10:08:35.189124643 +0000 UTC m=+2077.932670904" observedRunningTime="2025-12-10 10:08:35.857201626 +0000 UTC m=+2078.600747877" watchObservedRunningTime="2025-12-10 10:08:35.868634473 +0000 UTC m=+2078.612180734" Dec 10 10:08:47 crc kubenswrapper[4715]: I1210 10:08:47.714495 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:08:47 crc kubenswrapper[4715]: I1210 10:08:47.715188 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:09:17 crc kubenswrapper[4715]: I1210 10:09:17.714455 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:09:17 crc kubenswrapper[4715]: I1210 10:09:17.715011 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:09:47 crc kubenswrapper[4715]: I1210 10:09:47.713822 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:09:47 crc kubenswrapper[4715]: I1210 10:09:47.714326 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:09:47 crc kubenswrapper[4715]: I1210 10:09:47.714378 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 10:09:47 crc kubenswrapper[4715]: I1210 10:09:47.715206 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b2eee0605c28977d3538b338dcb2d691766610d0548ab30f1e772bdc0f6e9ba"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 10:09:47 crc kubenswrapper[4715]: I1210 10:09:47.715266 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://5b2eee0605c28977d3538b338dcb2d691766610d0548ab30f1e772bdc0f6e9ba" gracePeriod=600 Dec 10 10:09:47 crc kubenswrapper[4715]: I1210 10:09:47.997267 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="5b2eee0605c28977d3538b338dcb2d691766610d0548ab30f1e772bdc0f6e9ba" exitCode=0 Dec 10 10:09:47 crc kubenswrapper[4715]: I1210 10:09:47.997470 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"5b2eee0605c28977d3538b338dcb2d691766610d0548ab30f1e772bdc0f6e9ba"} Dec 10 10:09:47 crc kubenswrapper[4715]: I1210 10:09:47.997676 4715 scope.go:117] "RemoveContainer" containerID="206ae0dc13297871763c8927bbd3b1884fd686132dd736d6b15078556e3210a7" Dec 10 10:09:49 crc kubenswrapper[4715]: I1210 10:09:49.012952 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c"} Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.030343 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-86rpr"] Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.034544 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.039553 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-86rpr"] Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.137095 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-utilities\") pod \"community-operators-86rpr\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.138340 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6t8s\" (UniqueName: \"kubernetes.io/projected/94b67e19-ffd6-4207-afb2-0fc623327540-kube-api-access-j6t8s\") pod \"community-operators-86rpr\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.138413 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-catalog-content\") pod \"community-operators-86rpr\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.240217 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6t8s\" (UniqueName: \"kubernetes.io/projected/94b67e19-ffd6-4207-afb2-0fc623327540-kube-api-access-j6t8s\") pod \"community-operators-86rpr\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.240300 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-catalog-content\") pod \"community-operators-86rpr\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.241024 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-catalog-content\") pod \"community-operators-86rpr\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.241077 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-utilities\") pod \"community-operators-86rpr\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.241395 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-utilities\") pod \"community-operators-86rpr\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.267119 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6t8s\" (UniqueName: \"kubernetes.io/projected/94b67e19-ffd6-4207-afb2-0fc623327540-kube-api-access-j6t8s\") pod \"community-operators-86rpr\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.356626 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:20 crc kubenswrapper[4715]: I1210 10:10:20.960022 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-86rpr"] Dec 10 10:10:21 crc kubenswrapper[4715]: I1210 10:10:21.312060 4715 generic.go:334] "Generic (PLEG): container finished" podID="94b67e19-ffd6-4207-afb2-0fc623327540" containerID="969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12" exitCode=0 Dec 10 10:10:21 crc kubenswrapper[4715]: I1210 10:10:21.312149 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86rpr" event={"ID":"94b67e19-ffd6-4207-afb2-0fc623327540","Type":"ContainerDied","Data":"969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12"} Dec 10 10:10:21 crc kubenswrapper[4715]: I1210 10:10:21.312370 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86rpr" event={"ID":"94b67e19-ffd6-4207-afb2-0fc623327540","Type":"ContainerStarted","Data":"54ab4ec29b56f575802913225116b21409636627c163f55ce1fd512d432a5933"} Dec 10 10:10:21 crc kubenswrapper[4715]: I1210 10:10:21.313860 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 10:10:22 crc kubenswrapper[4715]: I1210 10:10:22.326219 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86rpr" event={"ID":"94b67e19-ffd6-4207-afb2-0fc623327540","Type":"ContainerStarted","Data":"3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6"} Dec 10 10:10:23 crc kubenswrapper[4715]: I1210 10:10:23.337151 4715 generic.go:334] "Generic (PLEG): container finished" podID="94b67e19-ffd6-4207-afb2-0fc623327540" containerID="3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6" exitCode=0 Dec 10 10:10:23 crc kubenswrapper[4715]: I1210 10:10:23.337234 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86rpr" event={"ID":"94b67e19-ffd6-4207-afb2-0fc623327540","Type":"ContainerDied","Data":"3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6"} Dec 10 10:10:24 crc kubenswrapper[4715]: I1210 10:10:24.348340 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86rpr" event={"ID":"94b67e19-ffd6-4207-afb2-0fc623327540","Type":"ContainerStarted","Data":"f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325"} Dec 10 10:10:24 crc kubenswrapper[4715]: I1210 10:10:24.375485 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-86rpr" podStartSLOduration=1.698328641 podStartE2EDuration="4.375467133s" podCreationTimestamp="2025-12-10 10:10:20 +0000 UTC" firstStartedPulling="2025-12-10 10:10:21.313618176 +0000 UTC m=+2184.057164427" lastFinishedPulling="2025-12-10 10:10:23.990756668 +0000 UTC m=+2186.734302919" observedRunningTime="2025-12-10 10:10:24.368763952 +0000 UTC m=+2187.112310223" watchObservedRunningTime="2025-12-10 10:10:24.375467133 +0000 UTC m=+2187.119013384" Dec 10 10:10:30 crc kubenswrapper[4715]: I1210 10:10:30.357238 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:30 crc kubenswrapper[4715]: I1210 10:10:30.357847 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:30 crc kubenswrapper[4715]: I1210 10:10:30.407299 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:30 crc kubenswrapper[4715]: I1210 10:10:30.467850 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:30 crc kubenswrapper[4715]: I1210 10:10:30.650402 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-86rpr"] Dec 10 10:10:32 crc kubenswrapper[4715]: I1210 10:10:32.422349 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-86rpr" podUID="94b67e19-ffd6-4207-afb2-0fc623327540" containerName="registry-server" containerID="cri-o://f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325" gracePeriod=2 Dec 10 10:10:32 crc kubenswrapper[4715]: I1210 10:10:32.944453 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.121490 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-catalog-content\") pod \"94b67e19-ffd6-4207-afb2-0fc623327540\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.121620 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6t8s\" (UniqueName: \"kubernetes.io/projected/94b67e19-ffd6-4207-afb2-0fc623327540-kube-api-access-j6t8s\") pod \"94b67e19-ffd6-4207-afb2-0fc623327540\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.121699 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-utilities\") pod \"94b67e19-ffd6-4207-afb2-0fc623327540\" (UID: \"94b67e19-ffd6-4207-afb2-0fc623327540\") " Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.123834 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-utilities" (OuterVolumeSpecName: "utilities") pod "94b67e19-ffd6-4207-afb2-0fc623327540" (UID: "94b67e19-ffd6-4207-afb2-0fc623327540"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.129470 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94b67e19-ffd6-4207-afb2-0fc623327540-kube-api-access-j6t8s" (OuterVolumeSpecName: "kube-api-access-j6t8s") pod "94b67e19-ffd6-4207-afb2-0fc623327540" (UID: "94b67e19-ffd6-4207-afb2-0fc623327540"). InnerVolumeSpecName "kube-api-access-j6t8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.172926 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94b67e19-ffd6-4207-afb2-0fc623327540" (UID: "94b67e19-ffd6-4207-afb2-0fc623327540"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.223878 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.224173 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6t8s\" (UniqueName: \"kubernetes.io/projected/94b67e19-ffd6-4207-afb2-0fc623327540-kube-api-access-j6t8s\") on node \"crc\" DevicePath \"\"" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.224186 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94b67e19-ffd6-4207-afb2-0fc623327540-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.434737 4715 generic.go:334] "Generic (PLEG): container finished" podID="94b67e19-ffd6-4207-afb2-0fc623327540" containerID="f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325" exitCode=0 Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.434868 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86rpr" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.435696 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86rpr" event={"ID":"94b67e19-ffd6-4207-afb2-0fc623327540","Type":"ContainerDied","Data":"f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325"} Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.435825 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86rpr" event={"ID":"94b67e19-ffd6-4207-afb2-0fc623327540","Type":"ContainerDied","Data":"54ab4ec29b56f575802913225116b21409636627c163f55ce1fd512d432a5933"} Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.435944 4715 scope.go:117] "RemoveContainer" containerID="f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.475822 4715 scope.go:117] "RemoveContainer" containerID="3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.482018 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-86rpr"] Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.489485 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-86rpr"] Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.515351 4715 scope.go:117] "RemoveContainer" containerID="969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.548829 4715 scope.go:117] "RemoveContainer" containerID="f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325" Dec 10 10:10:33 crc kubenswrapper[4715]: E1210 10:10:33.549475 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325\": container with ID starting with f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325 not found: ID does not exist" containerID="f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.549529 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325"} err="failed to get container status \"f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325\": rpc error: code = NotFound desc = could not find container \"f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325\": container with ID starting with f4c8938d0d4f14ab75364ca220734c65621839ba0447d1e04c1dcad72fca8325 not found: ID does not exist" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.549565 4715 scope.go:117] "RemoveContainer" containerID="3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6" Dec 10 10:10:33 crc kubenswrapper[4715]: E1210 10:10:33.550082 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6\": container with ID starting with 3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6 not found: ID does not exist" containerID="3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.550116 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6"} err="failed to get container status \"3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6\": rpc error: code = NotFound desc = could not find container \"3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6\": container with ID starting with 3d49d5b417212670300c13e3a052368a9c7c914c1a4c4c80907bfe7f7ebf39d6 not found: ID does not exist" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.550140 4715 scope.go:117] "RemoveContainer" containerID="969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12" Dec 10 10:10:33 crc kubenswrapper[4715]: E1210 10:10:33.550492 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12\": container with ID starting with 969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12 not found: ID does not exist" containerID="969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.550529 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12"} err="failed to get container status \"969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12\": rpc error: code = NotFound desc = could not find container \"969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12\": container with ID starting with 969138dbdb9ed830308c8b672c3bf6f3894b87bc8cd1734d10dce1c44a247c12 not found: ID does not exist" Dec 10 10:10:33 crc kubenswrapper[4715]: I1210 10:10:33.623612 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94b67e19-ffd6-4207-afb2-0fc623327540" path="/var/lib/kubelet/pods/94b67e19-ffd6-4207-afb2-0fc623327540/volumes" Dec 10 10:12:17 crc kubenswrapper[4715]: I1210 10:12:17.714423 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:12:17 crc kubenswrapper[4715]: I1210 10:12:17.715599 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.239543 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwv8"] Dec 10 10:12:18 crc kubenswrapper[4715]: E1210 10:12:18.240164 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b67e19-ffd6-4207-afb2-0fc623327540" containerName="extract-content" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.240189 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b67e19-ffd6-4207-afb2-0fc623327540" containerName="extract-content" Dec 10 10:12:18 crc kubenswrapper[4715]: E1210 10:12:18.240606 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b67e19-ffd6-4207-afb2-0fc623327540" containerName="registry-server" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.240620 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b67e19-ffd6-4207-afb2-0fc623327540" containerName="registry-server" Dec 10 10:12:18 crc kubenswrapper[4715]: E1210 10:12:18.240645 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b67e19-ffd6-4207-afb2-0fc623327540" containerName="extract-utilities" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.240653 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b67e19-ffd6-4207-afb2-0fc623327540" containerName="extract-utilities" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.240874 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="94b67e19-ffd6-4207-afb2-0fc623327540" containerName="registry-server" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.242807 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.257576 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwv8"] Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.353884 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-utilities\") pod \"redhat-marketplace-nxwv8\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.353996 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jrhz\" (UniqueName: \"kubernetes.io/projected/c693189e-c483-4b82-a3de-3fdb986e38d2-kube-api-access-5jrhz\") pod \"redhat-marketplace-nxwv8\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.354187 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-catalog-content\") pod \"redhat-marketplace-nxwv8\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.455836 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-utilities\") pod \"redhat-marketplace-nxwv8\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.455930 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jrhz\" (UniqueName: \"kubernetes.io/projected/c693189e-c483-4b82-a3de-3fdb986e38d2-kube-api-access-5jrhz\") pod \"redhat-marketplace-nxwv8\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.456042 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-catalog-content\") pod \"redhat-marketplace-nxwv8\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.456596 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-catalog-content\") pod \"redhat-marketplace-nxwv8\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.456597 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-utilities\") pod \"redhat-marketplace-nxwv8\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.479334 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jrhz\" (UniqueName: \"kubernetes.io/projected/c693189e-c483-4b82-a3de-3fdb986e38d2-kube-api-access-5jrhz\") pod \"redhat-marketplace-nxwv8\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:18 crc kubenswrapper[4715]: I1210 10:12:18.572568 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:19 crc kubenswrapper[4715]: I1210 10:12:19.086388 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwv8"] Dec 10 10:12:19 crc kubenswrapper[4715]: I1210 10:12:19.535326 4715 generic.go:334] "Generic (PLEG): container finished" podID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerID="c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886" exitCode=0 Dec 10 10:12:19 crc kubenswrapper[4715]: I1210 10:12:19.535372 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwv8" event={"ID":"c693189e-c483-4b82-a3de-3fdb986e38d2","Type":"ContainerDied","Data":"c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886"} Dec 10 10:12:19 crc kubenswrapper[4715]: I1210 10:12:19.535418 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwv8" event={"ID":"c693189e-c483-4b82-a3de-3fdb986e38d2","Type":"ContainerStarted","Data":"e4e4e7201ff822bf45cb9a641c4dbe353f06fa78aa00c35d8ed9a7e93d32740f"} Dec 10 10:12:22 crc kubenswrapper[4715]: I1210 10:12:22.565659 4715 generic.go:334] "Generic (PLEG): container finished" podID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerID="4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010" exitCode=0 Dec 10 10:12:22 crc kubenswrapper[4715]: I1210 10:12:22.565766 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwv8" event={"ID":"c693189e-c483-4b82-a3de-3fdb986e38d2","Type":"ContainerDied","Data":"4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010"} Dec 10 10:12:23 crc kubenswrapper[4715]: I1210 10:12:23.581398 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwv8" event={"ID":"c693189e-c483-4b82-a3de-3fdb986e38d2","Type":"ContainerStarted","Data":"f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690"} Dec 10 10:12:23 crc kubenswrapper[4715]: I1210 10:12:23.605494 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nxwv8" podStartSLOduration=2.15547816 podStartE2EDuration="5.605469861s" podCreationTimestamp="2025-12-10 10:12:18 +0000 UTC" firstStartedPulling="2025-12-10 10:12:19.537293743 +0000 UTC m=+2302.280839994" lastFinishedPulling="2025-12-10 10:12:22.987285444 +0000 UTC m=+2305.730831695" observedRunningTime="2025-12-10 10:12:23.596241796 +0000 UTC m=+2306.339788047" watchObservedRunningTime="2025-12-10 10:12:23.605469861 +0000 UTC m=+2306.349016112" Dec 10 10:12:28 crc kubenswrapper[4715]: I1210 10:12:28.574215 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:28 crc kubenswrapper[4715]: I1210 10:12:28.576114 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:28 crc kubenswrapper[4715]: I1210 10:12:28.630724 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:29 crc kubenswrapper[4715]: I1210 10:12:29.698871 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:30 crc kubenswrapper[4715]: I1210 10:12:30.023410 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwv8"] Dec 10 10:12:31 crc kubenswrapper[4715]: I1210 10:12:31.657475 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nxwv8" podUID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerName="registry-server" containerID="cri-o://f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690" gracePeriod=2 Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.074427 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.241863 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jrhz\" (UniqueName: \"kubernetes.io/projected/c693189e-c483-4b82-a3de-3fdb986e38d2-kube-api-access-5jrhz\") pod \"c693189e-c483-4b82-a3de-3fdb986e38d2\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.242121 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-utilities\") pod \"c693189e-c483-4b82-a3de-3fdb986e38d2\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.242249 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-catalog-content\") pod \"c693189e-c483-4b82-a3de-3fdb986e38d2\" (UID: \"c693189e-c483-4b82-a3de-3fdb986e38d2\") " Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.243225 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-utilities" (OuterVolumeSpecName: "utilities") pod "c693189e-c483-4b82-a3de-3fdb986e38d2" (UID: "c693189e-c483-4b82-a3de-3fdb986e38d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.248609 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c693189e-c483-4b82-a3de-3fdb986e38d2-kube-api-access-5jrhz" (OuterVolumeSpecName: "kube-api-access-5jrhz") pod "c693189e-c483-4b82-a3de-3fdb986e38d2" (UID: "c693189e-c483-4b82-a3de-3fdb986e38d2"). InnerVolumeSpecName "kube-api-access-5jrhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.255904 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.255958 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jrhz\" (UniqueName: \"kubernetes.io/projected/c693189e-c483-4b82-a3de-3fdb986e38d2-kube-api-access-5jrhz\") on node \"crc\" DevicePath \"\"" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.272116 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c693189e-c483-4b82-a3de-3fdb986e38d2" (UID: "c693189e-c483-4b82-a3de-3fdb986e38d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.358222 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c693189e-c483-4b82-a3de-3fdb986e38d2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.671072 4715 generic.go:334] "Generic (PLEG): container finished" podID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerID="f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690" exitCode=0 Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.671145 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxwv8" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.671138 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwv8" event={"ID":"c693189e-c483-4b82-a3de-3fdb986e38d2","Type":"ContainerDied","Data":"f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690"} Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.671573 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwv8" event={"ID":"c693189e-c483-4b82-a3de-3fdb986e38d2","Type":"ContainerDied","Data":"e4e4e7201ff822bf45cb9a641c4dbe353f06fa78aa00c35d8ed9a7e93d32740f"} Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.671596 4715 scope.go:117] "RemoveContainer" containerID="f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.706140 4715 scope.go:117] "RemoveContainer" containerID="4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.715283 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwv8"] Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.725695 4715 scope.go:117] "RemoveContainer" containerID="c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.728015 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwv8"] Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.778253 4715 scope.go:117] "RemoveContainer" containerID="f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690" Dec 10 10:12:32 crc kubenswrapper[4715]: E1210 10:12:32.778665 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690\": container with ID starting with f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690 not found: ID does not exist" containerID="f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.778696 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690"} err="failed to get container status \"f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690\": rpc error: code = NotFound desc = could not find container \"f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690\": container with ID starting with f868758571023610f201fb477cad59d741fc72acce1985c9fb33ae77c5ba3690 not found: ID does not exist" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.778718 4715 scope.go:117] "RemoveContainer" containerID="4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010" Dec 10 10:12:32 crc kubenswrapper[4715]: E1210 10:12:32.779193 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010\": container with ID starting with 4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010 not found: ID does not exist" containerID="4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.779214 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010"} err="failed to get container status \"4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010\": rpc error: code = NotFound desc = could not find container \"4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010\": container with ID starting with 4f6629204ca402ac3396d386245e235dc58ae40fe2e367b4b5b8e5a453fa1010 not found: ID does not exist" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.779228 4715 scope.go:117] "RemoveContainer" containerID="c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886" Dec 10 10:12:32 crc kubenswrapper[4715]: E1210 10:12:32.779485 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886\": container with ID starting with c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886 not found: ID does not exist" containerID="c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886" Dec 10 10:12:32 crc kubenswrapper[4715]: I1210 10:12:32.779622 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886"} err="failed to get container status \"c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886\": rpc error: code = NotFound desc = could not find container \"c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886\": container with ID starting with c12937135e676bde5094944c33ce671cebaea81f5c8fce6ba763a465c6a64886 not found: ID does not exist" Dec 10 10:12:33 crc kubenswrapper[4715]: I1210 10:12:33.615889 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c693189e-c483-4b82-a3de-3fdb986e38d2" path="/var/lib/kubelet/pods/c693189e-c483-4b82-a3de-3fdb986e38d2/volumes" Dec 10 10:12:47 crc kubenswrapper[4715]: I1210 10:12:47.714333 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:12:47 crc kubenswrapper[4715]: I1210 10:12:47.714807 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:12:52 crc kubenswrapper[4715]: I1210 10:12:52.875181 4715 generic.go:334] "Generic (PLEG): container finished" podID="cd86f221-0c52-4565-87ed-9d4c473aaaf4" containerID="6e7a58fc5a75316e2ef9ddc59edc4e5309cb863dbae332b492d19af73fd45b73" exitCode=0 Dec 10 10:12:52 crc kubenswrapper[4715]: I1210 10:12:52.875837 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" event={"ID":"cd86f221-0c52-4565-87ed-9d4c473aaaf4","Type":"ContainerDied","Data":"6e7a58fc5a75316e2ef9ddc59edc4e5309cb863dbae332b492d19af73fd45b73"} Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.272178 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.423407 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-inventory\") pod \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.423456 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-combined-ca-bundle\") pod \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.423670 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-ssh-key\") pod \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.423691 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-secret-0\") pod \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.423738 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sckcq\" (UniqueName: \"kubernetes.io/projected/cd86f221-0c52-4565-87ed-9d4c473aaaf4-kube-api-access-sckcq\") pod \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\" (UID: \"cd86f221-0c52-4565-87ed-9d4c473aaaf4\") " Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.429434 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "cd86f221-0c52-4565-87ed-9d4c473aaaf4" (UID: "cd86f221-0c52-4565-87ed-9d4c473aaaf4"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.430135 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd86f221-0c52-4565-87ed-9d4c473aaaf4-kube-api-access-sckcq" (OuterVolumeSpecName: "kube-api-access-sckcq") pod "cd86f221-0c52-4565-87ed-9d4c473aaaf4" (UID: "cd86f221-0c52-4565-87ed-9d4c473aaaf4"). InnerVolumeSpecName "kube-api-access-sckcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.462691 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-inventory" (OuterVolumeSpecName: "inventory") pod "cd86f221-0c52-4565-87ed-9d4c473aaaf4" (UID: "cd86f221-0c52-4565-87ed-9d4c473aaaf4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.473133 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "cd86f221-0c52-4565-87ed-9d4c473aaaf4" (UID: "cd86f221-0c52-4565-87ed-9d4c473aaaf4"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.480861 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cd86f221-0c52-4565-87ed-9d4c473aaaf4" (UID: "cd86f221-0c52-4565-87ed-9d4c473aaaf4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.525873 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.526127 4715 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.526202 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.526264 4715 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/cd86f221-0c52-4565-87ed-9d4c473aaaf4-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.526364 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sckcq\" (UniqueName: \"kubernetes.io/projected/cd86f221-0c52-4565-87ed-9d4c473aaaf4-kube-api-access-sckcq\") on node \"crc\" DevicePath \"\"" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.891633 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" event={"ID":"cd86f221-0c52-4565-87ed-9d4c473aaaf4","Type":"ContainerDied","Data":"733041d536eabd55811ac98a8a1a3fea750f11568fa6cedc996d7137591b6f06"} Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.891681 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="733041d536eabd55811ac98a8a1a3fea750f11568fa6cedc996d7137591b6f06" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.891731 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-959dc" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.994691 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp"] Dec 10 10:12:54 crc kubenswrapper[4715]: E1210 10:12:54.995785 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd86f221-0c52-4565-87ed-9d4c473aaaf4" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.995806 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd86f221-0c52-4565-87ed-9d4c473aaaf4" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 10:12:54 crc kubenswrapper[4715]: E1210 10:12:54.996051 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerName="extract-utilities" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.996064 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerName="extract-utilities" Dec 10 10:12:54 crc kubenswrapper[4715]: E1210 10:12:54.996201 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerName="registry-server" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.997874 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerName="registry-server" Dec 10 10:12:54 crc kubenswrapper[4715]: E1210 10:12:54.998070 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerName="extract-content" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.998085 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerName="extract-content" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.998454 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd86f221-0c52-4565-87ed-9d4c473aaaf4" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.998480 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c693189e-c483-4b82-a3de-3fdb986e38d2" containerName="registry-server" Dec 10 10:12:54 crc kubenswrapper[4715]: I1210 10:12:54.999383 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.003546 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.004077 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.004301 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.004506 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.004696 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.005056 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.005304 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.011467 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp"] Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.139273 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmv5n\" (UniqueName: \"kubernetes.io/projected/31ba5b33-f7ca-4389-879d-8a6a38cc0795-kube-api-access-qmv5n\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.139822 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.139882 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.140035 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.140151 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.140229 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.140310 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.140398 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.140503 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.241550 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmv5n\" (UniqueName: \"kubernetes.io/projected/31ba5b33-f7ca-4389-879d-8a6a38cc0795-kube-api-access-qmv5n\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.241622 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.241651 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.241690 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.241745 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.242761 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.243073 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.243146 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.243529 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.244309 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.249028 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.249028 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.249271 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.249654 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.249976 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.250254 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.253269 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.257794 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmv5n\" (UniqueName: \"kubernetes.io/projected/31ba5b33-f7ca-4389-879d-8a6a38cc0795-kube-api-access-qmv5n\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sbpnp\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.320533 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.827477 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp"] Dec 10 10:12:55 crc kubenswrapper[4715]: I1210 10:12:55.901256 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" event={"ID":"31ba5b33-f7ca-4389-879d-8a6a38cc0795","Type":"ContainerStarted","Data":"6bfbb176b441f2fb0ef5f1655dd14b916b62eeca93ac8ce1d3667dc9fa12e896"} Dec 10 10:12:56 crc kubenswrapper[4715]: I1210 10:12:56.912325 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" event={"ID":"31ba5b33-f7ca-4389-879d-8a6a38cc0795","Type":"ContainerStarted","Data":"41f1697f1ac44cfba119309b4434df9c0c25d02b207fa175d57848178784057a"} Dec 10 10:12:56 crc kubenswrapper[4715]: I1210 10:12:56.932261 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" podStartSLOduration=2.149195071 podStartE2EDuration="2.93223981s" podCreationTimestamp="2025-12-10 10:12:54 +0000 UTC" firstStartedPulling="2025-12-10 10:12:55.830958905 +0000 UTC m=+2338.574505156" lastFinishedPulling="2025-12-10 10:12:56.614003644 +0000 UTC m=+2339.357549895" observedRunningTime="2025-12-10 10:12:56.926645309 +0000 UTC m=+2339.670191560" watchObservedRunningTime="2025-12-10 10:12:56.93223981 +0000 UTC m=+2339.675786061" Dec 10 10:13:17 crc kubenswrapper[4715]: I1210 10:13:17.714434 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:13:17 crc kubenswrapper[4715]: I1210 10:13:17.714970 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:13:17 crc kubenswrapper[4715]: I1210 10:13:17.715036 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 10:13:17 crc kubenswrapper[4715]: I1210 10:13:17.715630 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 10:13:17 crc kubenswrapper[4715]: I1210 10:13:17.715782 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" gracePeriod=600 Dec 10 10:13:17 crc kubenswrapper[4715]: E1210 10:13:17.837344 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:13:18 crc kubenswrapper[4715]: I1210 10:13:18.130665 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" exitCode=0 Dec 10 10:13:18 crc kubenswrapper[4715]: I1210 10:13:18.130713 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c"} Dec 10 10:13:18 crc kubenswrapper[4715]: I1210 10:13:18.130763 4715 scope.go:117] "RemoveContainer" containerID="5b2eee0605c28977d3538b338dcb2d691766610d0548ab30f1e772bdc0f6e9ba" Dec 10 10:13:18 crc kubenswrapper[4715]: I1210 10:13:18.131517 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:13:18 crc kubenswrapper[4715]: E1210 10:13:18.131811 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:13:30 crc kubenswrapper[4715]: I1210 10:13:30.605517 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:13:30 crc kubenswrapper[4715]: E1210 10:13:30.606359 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:13:41 crc kubenswrapper[4715]: I1210 10:13:41.605258 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:13:41 crc kubenswrapper[4715]: E1210 10:13:41.606761 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:13:53 crc kubenswrapper[4715]: I1210 10:13:53.605413 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:13:53 crc kubenswrapper[4715]: E1210 10:13:53.606334 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:14:08 crc kubenswrapper[4715]: I1210 10:14:08.605871 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:14:08 crc kubenswrapper[4715]: E1210 10:14:08.606876 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:14:20 crc kubenswrapper[4715]: I1210 10:14:20.604870 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:14:20 crc kubenswrapper[4715]: E1210 10:14:20.605716 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:14:31 crc kubenswrapper[4715]: I1210 10:14:31.604988 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:14:31 crc kubenswrapper[4715]: E1210 10:14:31.605897 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.660133 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q2t9s"] Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.664530 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.673791 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q2t9s"] Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.816898 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-utilities\") pod \"redhat-operators-q2t9s\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.817130 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88h7v\" (UniqueName: \"kubernetes.io/projected/20587dbb-31fa-4741-a177-2185501510dc-kube-api-access-88h7v\") pod \"redhat-operators-q2t9s\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.817242 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-catalog-content\") pod \"redhat-operators-q2t9s\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.919250 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-utilities\") pod \"redhat-operators-q2t9s\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.919654 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88h7v\" (UniqueName: \"kubernetes.io/projected/20587dbb-31fa-4741-a177-2185501510dc-kube-api-access-88h7v\") pod \"redhat-operators-q2t9s\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.919694 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-catalog-content\") pod \"redhat-operators-q2t9s\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.920092 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-utilities\") pod \"redhat-operators-q2t9s\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.920210 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-catalog-content\") pod \"redhat-operators-q2t9s\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.944155 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88h7v\" (UniqueName: \"kubernetes.io/projected/20587dbb-31fa-4741-a177-2185501510dc-kube-api-access-88h7v\") pod \"redhat-operators-q2t9s\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:36 crc kubenswrapper[4715]: I1210 10:14:36.994246 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:37 crc kubenswrapper[4715]: I1210 10:14:37.499020 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q2t9s"] Dec 10 10:14:37 crc kubenswrapper[4715]: I1210 10:14:37.954002 4715 generic.go:334] "Generic (PLEG): container finished" podID="20587dbb-31fa-4741-a177-2185501510dc" containerID="54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3" exitCode=0 Dec 10 10:14:37 crc kubenswrapper[4715]: I1210 10:14:37.954239 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t9s" event={"ID":"20587dbb-31fa-4741-a177-2185501510dc","Type":"ContainerDied","Data":"54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3"} Dec 10 10:14:37 crc kubenswrapper[4715]: I1210 10:14:37.954288 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t9s" event={"ID":"20587dbb-31fa-4741-a177-2185501510dc","Type":"ContainerStarted","Data":"c6db3d4db0d3f2f4bd7646384c4df21d73610d9140ee1119d1b1446bccc15fa2"} Dec 10 10:14:39 crc kubenswrapper[4715]: I1210 10:14:39.974271 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t9s" event={"ID":"20587dbb-31fa-4741-a177-2185501510dc","Type":"ContainerStarted","Data":"c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc"} Dec 10 10:14:42 crc kubenswrapper[4715]: I1210 10:14:42.000343 4715 generic.go:334] "Generic (PLEG): container finished" podID="20587dbb-31fa-4741-a177-2185501510dc" containerID="c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc" exitCode=0 Dec 10 10:14:42 crc kubenswrapper[4715]: I1210 10:14:42.000356 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t9s" event={"ID":"20587dbb-31fa-4741-a177-2185501510dc","Type":"ContainerDied","Data":"c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc"} Dec 10 10:14:42 crc kubenswrapper[4715]: I1210 10:14:42.605655 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:14:42 crc kubenswrapper[4715]: E1210 10:14:42.606094 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:14:44 crc kubenswrapper[4715]: I1210 10:14:44.023898 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t9s" event={"ID":"20587dbb-31fa-4741-a177-2185501510dc","Type":"ContainerStarted","Data":"4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832"} Dec 10 10:14:44 crc kubenswrapper[4715]: I1210 10:14:44.051181 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q2t9s" podStartSLOduration=2.505116806 podStartE2EDuration="8.051153529s" podCreationTimestamp="2025-12-10 10:14:36 +0000 UTC" firstStartedPulling="2025-12-10 10:14:37.955696043 +0000 UTC m=+2440.699242294" lastFinishedPulling="2025-12-10 10:14:43.501732766 +0000 UTC m=+2446.245279017" observedRunningTime="2025-12-10 10:14:44.041266905 +0000 UTC m=+2446.784813146" watchObservedRunningTime="2025-12-10 10:14:44.051153529 +0000 UTC m=+2446.794699780" Dec 10 10:14:46 crc kubenswrapper[4715]: I1210 10:14:46.994886 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:46 crc kubenswrapper[4715]: I1210 10:14:46.995266 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:48 crc kubenswrapper[4715]: I1210 10:14:48.042701 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q2t9s" podUID="20587dbb-31fa-4741-a177-2185501510dc" containerName="registry-server" probeResult="failure" output=< Dec 10 10:14:48 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 10 10:14:48 crc kubenswrapper[4715]: > Dec 10 10:14:54 crc kubenswrapper[4715]: I1210 10:14:54.605480 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:14:54 crc kubenswrapper[4715]: E1210 10:14:54.606532 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:14:57 crc kubenswrapper[4715]: I1210 10:14:57.041272 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:57 crc kubenswrapper[4715]: I1210 10:14:57.088871 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:57 crc kubenswrapper[4715]: I1210 10:14:57.281680 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q2t9s"] Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.141508 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q2t9s" podUID="20587dbb-31fa-4741-a177-2185501510dc" containerName="registry-server" containerID="cri-o://4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832" gracePeriod=2 Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.621133 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.764797 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-utilities\") pod \"20587dbb-31fa-4741-a177-2185501510dc\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.764845 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88h7v\" (UniqueName: \"kubernetes.io/projected/20587dbb-31fa-4741-a177-2185501510dc-kube-api-access-88h7v\") pod \"20587dbb-31fa-4741-a177-2185501510dc\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.764872 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-catalog-content\") pod \"20587dbb-31fa-4741-a177-2185501510dc\" (UID: \"20587dbb-31fa-4741-a177-2185501510dc\") " Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.765719 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-utilities" (OuterVolumeSpecName: "utilities") pod "20587dbb-31fa-4741-a177-2185501510dc" (UID: "20587dbb-31fa-4741-a177-2185501510dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.771268 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20587dbb-31fa-4741-a177-2185501510dc-kube-api-access-88h7v" (OuterVolumeSpecName: "kube-api-access-88h7v") pod "20587dbb-31fa-4741-a177-2185501510dc" (UID: "20587dbb-31fa-4741-a177-2185501510dc"). InnerVolumeSpecName "kube-api-access-88h7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.867743 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.867776 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88h7v\" (UniqueName: \"kubernetes.io/projected/20587dbb-31fa-4741-a177-2185501510dc-kube-api-access-88h7v\") on node \"crc\" DevicePath \"\"" Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.875304 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20587dbb-31fa-4741-a177-2185501510dc" (UID: "20587dbb-31fa-4741-a177-2185501510dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:14:58 crc kubenswrapper[4715]: I1210 10:14:58.968598 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20587dbb-31fa-4741-a177-2185501510dc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.154243 4715 generic.go:334] "Generic (PLEG): container finished" podID="20587dbb-31fa-4741-a177-2185501510dc" containerID="4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832" exitCode=0 Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.154307 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t9s" event={"ID":"20587dbb-31fa-4741-a177-2185501510dc","Type":"ContainerDied","Data":"4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832"} Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.154322 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2t9s" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.154334 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2t9s" event={"ID":"20587dbb-31fa-4741-a177-2185501510dc","Type":"ContainerDied","Data":"c6db3d4db0d3f2f4bd7646384c4df21d73610d9140ee1119d1b1446bccc15fa2"} Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.154353 4715 scope.go:117] "RemoveContainer" containerID="4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.201741 4715 scope.go:117] "RemoveContainer" containerID="c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.217024 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q2t9s"] Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.228251 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q2t9s"] Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.229181 4715 scope.go:117] "RemoveContainer" containerID="54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.266188 4715 scope.go:117] "RemoveContainer" containerID="4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832" Dec 10 10:14:59 crc kubenswrapper[4715]: E1210 10:14:59.266624 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832\": container with ID starting with 4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832 not found: ID does not exist" containerID="4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.266659 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832"} err="failed to get container status \"4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832\": rpc error: code = NotFound desc = could not find container \"4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832\": container with ID starting with 4a5911b0c5d3d98d3725de19f9169802557738356bcded5ce0184ad86a70b832 not found: ID does not exist" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.266682 4715 scope.go:117] "RemoveContainer" containerID="c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc" Dec 10 10:14:59 crc kubenswrapper[4715]: E1210 10:14:59.266945 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc\": container with ID starting with c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc not found: ID does not exist" containerID="c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.267061 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc"} err="failed to get container status \"c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc\": rpc error: code = NotFound desc = could not find container \"c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc\": container with ID starting with c399587964b0aaf0157047b630f4b2519b11b299835186d83b66696b55db80cc not found: ID does not exist" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.267159 4715 scope.go:117] "RemoveContainer" containerID="54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3" Dec 10 10:14:59 crc kubenswrapper[4715]: E1210 10:14:59.267538 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3\": container with ID starting with 54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3 not found: ID does not exist" containerID="54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.267565 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3"} err="failed to get container status \"54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3\": rpc error: code = NotFound desc = could not find container \"54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3\": container with ID starting with 54ce7448f1ba0e12a08cc852493848e0a0923a629ae45f171be0ebfbccc08fb3 not found: ID does not exist" Dec 10 10:14:59 crc kubenswrapper[4715]: E1210 10:14:59.313352 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20587dbb_31fa_4741_a177_2185501510dc.slice/crio-c6db3d4db0d3f2f4bd7646384c4df21d73610d9140ee1119d1b1446bccc15fa2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20587dbb_31fa_4741_a177_2185501510dc.slice\": RecentStats: unable to find data in memory cache]" Dec 10 10:14:59 crc kubenswrapper[4715]: I1210 10:14:59.619582 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20587dbb-31fa-4741-a177-2185501510dc" path="/var/lib/kubelet/pods/20587dbb-31fa-4741-a177-2185501510dc/volumes" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.174162 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2"] Dec 10 10:15:00 crc kubenswrapper[4715]: E1210 10:15:00.174668 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20587dbb-31fa-4741-a177-2185501510dc" containerName="extract-utilities" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.174686 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="20587dbb-31fa-4741-a177-2185501510dc" containerName="extract-utilities" Dec 10 10:15:00 crc kubenswrapper[4715]: E1210 10:15:00.174717 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20587dbb-31fa-4741-a177-2185501510dc" containerName="registry-server" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.174723 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="20587dbb-31fa-4741-a177-2185501510dc" containerName="registry-server" Dec 10 10:15:00 crc kubenswrapper[4715]: E1210 10:15:00.174741 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20587dbb-31fa-4741-a177-2185501510dc" containerName="extract-content" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.174747 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="20587dbb-31fa-4741-a177-2185501510dc" containerName="extract-content" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.175030 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="20587dbb-31fa-4741-a177-2185501510dc" containerName="registry-server" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.175642 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.180690 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.187267 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.191297 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2"] Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.304027 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-secret-volume\") pod \"collect-profiles-29422695-zpqd2\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.304979 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-config-volume\") pod \"collect-profiles-29422695-zpqd2\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.305518 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s96nt\" (UniqueName: \"kubernetes.io/projected/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-kube-api-access-s96nt\") pod \"collect-profiles-29422695-zpqd2\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.408532 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s96nt\" (UniqueName: \"kubernetes.io/projected/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-kube-api-access-s96nt\") pod \"collect-profiles-29422695-zpqd2\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.408687 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-secret-volume\") pod \"collect-profiles-29422695-zpqd2\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.408736 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-config-volume\") pod \"collect-profiles-29422695-zpqd2\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.409750 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-config-volume\") pod \"collect-profiles-29422695-zpqd2\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.415606 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-secret-volume\") pod \"collect-profiles-29422695-zpqd2\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.426396 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s96nt\" (UniqueName: \"kubernetes.io/projected/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-kube-api-access-s96nt\") pod \"collect-profiles-29422695-zpqd2\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.496623 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:00 crc kubenswrapper[4715]: I1210 10:15:00.962253 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2"] Dec 10 10:15:01 crc kubenswrapper[4715]: I1210 10:15:01.180758 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" event={"ID":"5255a0f9-a6ec-4ed2-bb79-97b4761128ec","Type":"ContainerStarted","Data":"da522ee59194a471bd8dde7e75edf2b0ffa0b4543654bd699cf6ce71c84cb6dc"} Dec 10 10:15:01 crc kubenswrapper[4715]: I1210 10:15:01.180804 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" event={"ID":"5255a0f9-a6ec-4ed2-bb79-97b4761128ec","Type":"ContainerStarted","Data":"80657a300813902eb96925dcb5b77a16243389faa6ace4a86906d218ca0e1ea2"} Dec 10 10:15:01 crc kubenswrapper[4715]: I1210 10:15:01.200398 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" podStartSLOduration=1.200376435 podStartE2EDuration="1.200376435s" podCreationTimestamp="2025-12-10 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 10:15:01.195531266 +0000 UTC m=+2463.939077517" watchObservedRunningTime="2025-12-10 10:15:01.200376435 +0000 UTC m=+2463.943922686" Dec 10 10:15:02 crc kubenswrapper[4715]: I1210 10:15:02.192208 4715 generic.go:334] "Generic (PLEG): container finished" podID="5255a0f9-a6ec-4ed2-bb79-97b4761128ec" containerID="da522ee59194a471bd8dde7e75edf2b0ffa0b4543654bd699cf6ce71c84cb6dc" exitCode=0 Dec 10 10:15:02 crc kubenswrapper[4715]: I1210 10:15:02.192268 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" event={"ID":"5255a0f9-a6ec-4ed2-bb79-97b4761128ec","Type":"ContainerDied","Data":"da522ee59194a471bd8dde7e75edf2b0ffa0b4543654bd699cf6ce71c84cb6dc"} Dec 10 10:15:03 crc kubenswrapper[4715]: I1210 10:15:03.550387 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:03 crc kubenswrapper[4715]: I1210 10:15:03.692369 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-config-volume\") pod \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " Dec 10 10:15:03 crc kubenswrapper[4715]: I1210 10:15:03.692702 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-secret-volume\") pod \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " Dec 10 10:15:03 crc kubenswrapper[4715]: I1210 10:15:03.692849 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s96nt\" (UniqueName: \"kubernetes.io/projected/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-kube-api-access-s96nt\") pod \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\" (UID: \"5255a0f9-a6ec-4ed2-bb79-97b4761128ec\") " Dec 10 10:15:03 crc kubenswrapper[4715]: I1210 10:15:03.693866 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-config-volume" (OuterVolumeSpecName: "config-volume") pod "5255a0f9-a6ec-4ed2-bb79-97b4761128ec" (UID: "5255a0f9-a6ec-4ed2-bb79-97b4761128ec"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 10:15:03 crc kubenswrapper[4715]: I1210 10:15:03.694602 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:03 crc kubenswrapper[4715]: I1210 10:15:03.701827 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5255a0f9-a6ec-4ed2-bb79-97b4761128ec" (UID: "5255a0f9-a6ec-4ed2-bb79-97b4761128ec"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:15:03 crc kubenswrapper[4715]: I1210 10:15:03.701955 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-kube-api-access-s96nt" (OuterVolumeSpecName: "kube-api-access-s96nt") pod "5255a0f9-a6ec-4ed2-bb79-97b4761128ec" (UID: "5255a0f9-a6ec-4ed2-bb79-97b4761128ec"). InnerVolumeSpecName "kube-api-access-s96nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:15:03 crc kubenswrapper[4715]: I1210 10:15:03.796836 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:03 crc kubenswrapper[4715]: I1210 10:15:03.796883 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s96nt\" (UniqueName: \"kubernetes.io/projected/5255a0f9-a6ec-4ed2-bb79-97b4761128ec-kube-api-access-s96nt\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:04 crc kubenswrapper[4715]: I1210 10:15:04.211186 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" event={"ID":"5255a0f9-a6ec-4ed2-bb79-97b4761128ec","Type":"ContainerDied","Data":"80657a300813902eb96925dcb5b77a16243389faa6ace4a86906d218ca0e1ea2"} Dec 10 10:15:04 crc kubenswrapper[4715]: I1210 10:15:04.211234 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80657a300813902eb96925dcb5b77a16243389faa6ace4a86906d218ca0e1ea2" Dec 10 10:15:04 crc kubenswrapper[4715]: I1210 10:15:04.211241 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422695-zpqd2" Dec 10 10:15:04 crc kubenswrapper[4715]: I1210 10:15:04.273263 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975"] Dec 10 10:15:04 crc kubenswrapper[4715]: I1210 10:15:04.281307 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422650-8r975"] Dec 10 10:15:05 crc kubenswrapper[4715]: I1210 10:15:05.620165 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efb368e0-756b-4323-8ca2-0852ff385eb4" path="/var/lib/kubelet/pods/efb368e0-756b-4323-8ca2-0852ff385eb4/volumes" Dec 10 10:15:06 crc kubenswrapper[4715]: I1210 10:15:06.605457 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:15:06 crc kubenswrapper[4715]: E1210 10:15:06.605992 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:15:20 crc kubenswrapper[4715]: I1210 10:15:20.605384 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:15:20 crc kubenswrapper[4715]: E1210 10:15:20.606210 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:15:23 crc kubenswrapper[4715]: I1210 10:15:23.841974 4715 scope.go:117] "RemoveContainer" containerID="ba038ccc002ae5e69cf5603e7af827400c078a0a8a5db976d6d1797f0bbdaa48" Dec 10 10:15:32 crc kubenswrapper[4715]: I1210 10:15:32.605044 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:15:32 crc kubenswrapper[4715]: E1210 10:15:32.605834 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:15:42 crc kubenswrapper[4715]: I1210 10:15:42.741498 4715 generic.go:334] "Generic (PLEG): container finished" podID="31ba5b33-f7ca-4389-879d-8a6a38cc0795" containerID="41f1697f1ac44cfba119309b4434df9c0c25d02b207fa175d57848178784057a" exitCode=0 Dec 10 10:15:42 crc kubenswrapper[4715]: I1210 10:15:42.741569 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" event={"ID":"31ba5b33-f7ca-4389-879d-8a6a38cc0795","Type":"ContainerDied","Data":"41f1697f1ac44cfba119309b4434df9c0c25d02b207fa175d57848178784057a"} Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.156412 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.251156 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-extra-config-0\") pod \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.251364 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-inventory\") pod \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.251420 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmv5n\" (UniqueName: \"kubernetes.io/projected/31ba5b33-f7ca-4389-879d-8a6a38cc0795-kube-api-access-qmv5n\") pod \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.251466 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-ssh-key\") pod \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.251508 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-0\") pod \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.251554 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-combined-ca-bundle\") pod \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.251610 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-0\") pod \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.251637 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-1\") pod \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.251744 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-1\") pod \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\" (UID: \"31ba5b33-f7ca-4389-879d-8a6a38cc0795\") " Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.276261 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "31ba5b33-f7ca-4389-879d-8a6a38cc0795" (UID: "31ba5b33-f7ca-4389-879d-8a6a38cc0795"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.276289 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31ba5b33-f7ca-4389-879d-8a6a38cc0795-kube-api-access-qmv5n" (OuterVolumeSpecName: "kube-api-access-qmv5n") pod "31ba5b33-f7ca-4389-879d-8a6a38cc0795" (UID: "31ba5b33-f7ca-4389-879d-8a6a38cc0795"). InnerVolumeSpecName "kube-api-access-qmv5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.297208 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-inventory" (OuterVolumeSpecName: "inventory") pod "31ba5b33-f7ca-4389-879d-8a6a38cc0795" (UID: "31ba5b33-f7ca-4389-879d-8a6a38cc0795"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.297248 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "31ba5b33-f7ca-4389-879d-8a6a38cc0795" (UID: "31ba5b33-f7ca-4389-879d-8a6a38cc0795"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.298583 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "31ba5b33-f7ca-4389-879d-8a6a38cc0795" (UID: "31ba5b33-f7ca-4389-879d-8a6a38cc0795"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.316186 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "31ba5b33-f7ca-4389-879d-8a6a38cc0795" (UID: "31ba5b33-f7ca-4389-879d-8a6a38cc0795"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.320466 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "31ba5b33-f7ca-4389-879d-8a6a38cc0795" (UID: "31ba5b33-f7ca-4389-879d-8a6a38cc0795"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.325557 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "31ba5b33-f7ca-4389-879d-8a6a38cc0795" (UID: "31ba5b33-f7ca-4389-879d-8a6a38cc0795"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.327729 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "31ba5b33-f7ca-4389-879d-8a6a38cc0795" (UID: "31ba5b33-f7ca-4389-879d-8a6a38cc0795"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.353660 4715 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.353708 4715 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.353722 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmv5n\" (UniqueName: \"kubernetes.io/projected/31ba5b33-f7ca-4389-879d-8a6a38cc0795-kube-api-access-qmv5n\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.353732 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.353741 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.353750 4715 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.353759 4715 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.353767 4715 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.353777 4715 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/31ba5b33-f7ca-4389-879d-8a6a38cc0795-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.604423 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:15:44 crc kubenswrapper[4715]: E1210 10:15:44.604645 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.774127 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" event={"ID":"31ba5b33-f7ca-4389-879d-8a6a38cc0795","Type":"ContainerDied","Data":"6bfbb176b441f2fb0ef5f1655dd14b916b62eeca93ac8ce1d3667dc9fa12e896"} Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.774178 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bfbb176b441f2fb0ef5f1655dd14b916b62eeca93ac8ce1d3667dc9fa12e896" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.774275 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sbpnp" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.859443 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk"] Dec 10 10:15:44 crc kubenswrapper[4715]: E1210 10:15:44.859809 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31ba5b33-f7ca-4389-879d-8a6a38cc0795" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.859826 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="31ba5b33-f7ca-4389-879d-8a6a38cc0795" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 10:15:44 crc kubenswrapper[4715]: E1210 10:15:44.859869 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5255a0f9-a6ec-4ed2-bb79-97b4761128ec" containerName="collect-profiles" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.859876 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5255a0f9-a6ec-4ed2-bb79-97b4761128ec" containerName="collect-profiles" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.860111 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5255a0f9-a6ec-4ed2-bb79-97b4761128ec" containerName="collect-profiles" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.860137 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="31ba5b33-f7ca-4389-879d-8a6a38cc0795" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.862160 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.868617 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.870450 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.870490 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxjsw" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.870588 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.870688 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.875571 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk"] Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.974994 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.975420 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.975453 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.975506 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rvnf\" (UniqueName: \"kubernetes.io/projected/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-kube-api-access-9rvnf\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.975601 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.975628 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:44 crc kubenswrapper[4715]: I1210 10:15:44.975648 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.095532 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.095592 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.095885 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.096304 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.096382 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.096403 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.096453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rvnf\" (UniqueName: \"kubernetes.io/projected/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-kube-api-access-9rvnf\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.099687 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.099714 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.110242 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.110694 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.111482 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.111997 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.119182 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rvnf\" (UniqueName: \"kubernetes.io/projected/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-kube-api-access-9rvnf\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.184380 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.722167 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk"] Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.727401 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 10:15:45 crc kubenswrapper[4715]: I1210 10:15:45.786037 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" event={"ID":"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f","Type":"ContainerStarted","Data":"2cd3a538753158d9a397e9149c9e201fb920bc21c85ba181ea7297157cf05a50"} Dec 10 10:15:46 crc kubenswrapper[4715]: I1210 10:15:46.796964 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" event={"ID":"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f","Type":"ContainerStarted","Data":"d45a2d74af0272a26fd542d0815ccd80df009ac4f4f8e0ac78eac908ff54bcc9"} Dec 10 10:15:46 crc kubenswrapper[4715]: I1210 10:15:46.816720 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" podStartSLOduration=2.240444476 podStartE2EDuration="2.816702302s" podCreationTimestamp="2025-12-10 10:15:44 +0000 UTC" firstStartedPulling="2025-12-10 10:15:45.727040577 +0000 UTC m=+2508.470586828" lastFinishedPulling="2025-12-10 10:15:46.303298403 +0000 UTC m=+2509.046844654" observedRunningTime="2025-12-10 10:15:46.814815948 +0000 UTC m=+2509.558362209" watchObservedRunningTime="2025-12-10 10:15:46.816702302 +0000 UTC m=+2509.560248553" Dec 10 10:15:55 crc kubenswrapper[4715]: I1210 10:15:55.604641 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:15:55 crc kubenswrapper[4715]: E1210 10:15:55.606294 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:16:09 crc kubenswrapper[4715]: I1210 10:16:09.605017 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:16:09 crc kubenswrapper[4715]: E1210 10:16:09.605779 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:16:20 crc kubenswrapper[4715]: I1210 10:16:20.605993 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:16:20 crc kubenswrapper[4715]: E1210 10:16:20.607252 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:16:31 crc kubenswrapper[4715]: I1210 10:16:31.604867 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:16:31 crc kubenswrapper[4715]: E1210 10:16:31.605778 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:16:44 crc kubenswrapper[4715]: I1210 10:16:44.605581 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:16:44 crc kubenswrapper[4715]: E1210 10:16:44.606569 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:16:55 crc kubenswrapper[4715]: I1210 10:16:55.606382 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:16:55 crc kubenswrapper[4715]: E1210 10:16:55.607286 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:17:10 crc kubenswrapper[4715]: I1210 10:17:10.606976 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:17:10 crc kubenswrapper[4715]: E1210 10:17:10.607854 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:17:25 crc kubenswrapper[4715]: I1210 10:17:25.605941 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:17:25 crc kubenswrapper[4715]: E1210 10:17:25.606727 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:17:39 crc kubenswrapper[4715]: I1210 10:17:39.605345 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:17:39 crc kubenswrapper[4715]: E1210 10:17:39.607235 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:17:51 crc kubenswrapper[4715]: I1210 10:17:51.608818 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:17:51 crc kubenswrapper[4715]: E1210 10:17:51.609740 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:18:04 crc kubenswrapper[4715]: I1210 10:18:04.606178 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:18:04 crc kubenswrapper[4715]: E1210 10:18:04.607148 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:18:05 crc kubenswrapper[4715]: I1210 10:18:05.187610 4715 generic.go:334] "Generic (PLEG): container finished" podID="54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" containerID="d45a2d74af0272a26fd542d0815ccd80df009ac4f4f8e0ac78eac908ff54bcc9" exitCode=0 Dec 10 10:18:05 crc kubenswrapper[4715]: I1210 10:18:05.187689 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" event={"ID":"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f","Type":"ContainerDied","Data":"d45a2d74af0272a26fd542d0815ccd80df009ac4f4f8e0ac78eac908ff54bcc9"} Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.613098 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.626825 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-1\") pod \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.626995 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-inventory\") pod \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.627034 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-2\") pod \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.627070 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-0\") pod \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.627136 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-telemetry-combined-ca-bundle\") pod \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.627197 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rvnf\" (UniqueName: \"kubernetes.io/projected/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-kube-api-access-9rvnf\") pod \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.627314 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ssh-key\") pod \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\" (UID: \"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f\") " Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.634688 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-kube-api-access-9rvnf" (OuterVolumeSpecName: "kube-api-access-9rvnf") pod "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" (UID: "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f"). InnerVolumeSpecName "kube-api-access-9rvnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.637367 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" (UID: "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.656463 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" (UID: "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.672195 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-inventory" (OuterVolumeSpecName: "inventory") pod "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" (UID: "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.672599 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" (UID: "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.675391 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" (UID: "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.675903 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" (UID: "54173a4c-eb6d-4482-9bce-3f24ab7a8b4f"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.731248 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rvnf\" (UniqueName: \"kubernetes.io/projected/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-kube-api-access-9rvnf\") on node \"crc\" DevicePath \"\"" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.731304 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.731319 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.731334 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.731350 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.731363 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 10 10:18:06 crc kubenswrapper[4715]: I1210 10:18:06.731374 4715 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54173a4c-eb6d-4482-9bce-3f24ab7a8b4f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 10:18:07 crc kubenswrapper[4715]: I1210 10:18:07.216561 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" event={"ID":"54173a4c-eb6d-4482-9bce-3f24ab7a8b4f","Type":"ContainerDied","Data":"2cd3a538753158d9a397e9149c9e201fb920bc21c85ba181ea7297157cf05a50"} Dec 10 10:18:07 crc kubenswrapper[4715]: I1210 10:18:07.217063 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cd3a538753158d9a397e9149c9e201fb920bc21c85ba181ea7297157cf05a50" Dec 10 10:18:07 crc kubenswrapper[4715]: I1210 10:18:07.216605 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk" Dec 10 10:18:17 crc kubenswrapper[4715]: I1210 10:18:17.617520 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:18:17 crc kubenswrapper[4715]: E1210 10:18:17.618563 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.662965 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kw89m"] Dec 10 10:18:22 crc kubenswrapper[4715]: E1210 10:18:22.665016 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.665130 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.665459 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="54173a4c-eb6d-4482-9bce-3f24ab7a8b4f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.667058 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.677959 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kw89m"] Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.751478 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sszmf\" (UniqueName: \"kubernetes.io/projected/011917ba-84a4-4cc1-8a22-9e983eac49c1-kube-api-access-sszmf\") pod \"certified-operators-kw89m\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.751605 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-utilities\") pod \"certified-operators-kw89m\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.751653 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-catalog-content\") pod \"certified-operators-kw89m\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.854213 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-catalog-content\") pod \"certified-operators-kw89m\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.854359 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sszmf\" (UniqueName: \"kubernetes.io/projected/011917ba-84a4-4cc1-8a22-9e983eac49c1-kube-api-access-sszmf\") pod \"certified-operators-kw89m\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.854468 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-utilities\") pod \"certified-operators-kw89m\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.854977 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-utilities\") pod \"certified-operators-kw89m\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.856230 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-catalog-content\") pod \"certified-operators-kw89m\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:22 crc kubenswrapper[4715]: I1210 10:18:22.876869 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sszmf\" (UniqueName: \"kubernetes.io/projected/011917ba-84a4-4cc1-8a22-9e983eac49c1-kube-api-access-sszmf\") pod \"certified-operators-kw89m\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:23 crc kubenswrapper[4715]: I1210 10:18:23.004084 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:23 crc kubenswrapper[4715]: I1210 10:18:23.414489 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kw89m"] Dec 10 10:18:23 crc kubenswrapper[4715]: W1210 10:18:23.422880 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod011917ba_84a4_4cc1_8a22_9e983eac49c1.slice/crio-00ebffb0e66625cb08ab821c76a2964fb11598111e2e63c4619a8a2ba037b6ca WatchSource:0}: Error finding container 00ebffb0e66625cb08ab821c76a2964fb11598111e2e63c4619a8a2ba037b6ca: Status 404 returned error can't find the container with id 00ebffb0e66625cb08ab821c76a2964fb11598111e2e63c4619a8a2ba037b6ca Dec 10 10:18:24 crc kubenswrapper[4715]: I1210 10:18:24.429521 4715 generic.go:334] "Generic (PLEG): container finished" podID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerID="003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494" exitCode=0 Dec 10 10:18:24 crc kubenswrapper[4715]: I1210 10:18:24.430811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kw89m" event={"ID":"011917ba-84a4-4cc1-8a22-9e983eac49c1","Type":"ContainerDied","Data":"003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494"} Dec 10 10:18:24 crc kubenswrapper[4715]: I1210 10:18:24.430942 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kw89m" event={"ID":"011917ba-84a4-4cc1-8a22-9e983eac49c1","Type":"ContainerStarted","Data":"00ebffb0e66625cb08ab821c76a2964fb11598111e2e63c4619a8a2ba037b6ca"} Dec 10 10:18:26 crc kubenswrapper[4715]: I1210 10:18:26.452801 4715 generic.go:334] "Generic (PLEG): container finished" podID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerID="46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40" exitCode=0 Dec 10 10:18:26 crc kubenswrapper[4715]: I1210 10:18:26.452967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kw89m" event={"ID":"011917ba-84a4-4cc1-8a22-9e983eac49c1","Type":"ContainerDied","Data":"46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40"} Dec 10 10:18:27 crc kubenswrapper[4715]: I1210 10:18:27.466562 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kw89m" event={"ID":"011917ba-84a4-4cc1-8a22-9e983eac49c1","Type":"ContainerStarted","Data":"245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348"} Dec 10 10:18:27 crc kubenswrapper[4715]: I1210 10:18:27.490986 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kw89m" podStartSLOduration=2.913889506 podStartE2EDuration="5.49095817s" podCreationTimestamp="2025-12-10 10:18:22 +0000 UTC" firstStartedPulling="2025-12-10 10:18:24.435175748 +0000 UTC m=+2667.178721989" lastFinishedPulling="2025-12-10 10:18:27.012244402 +0000 UTC m=+2669.755790653" observedRunningTime="2025-12-10 10:18:27.485326399 +0000 UTC m=+2670.228872660" watchObservedRunningTime="2025-12-10 10:18:27.49095817 +0000 UTC m=+2670.234504421" Dec 10 10:18:30 crc kubenswrapper[4715]: I1210 10:18:30.605437 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:18:31 crc kubenswrapper[4715]: I1210 10:18:31.514669 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"ac70cfef39940dc73d054eb086c3ad80da75d6389147830f13b583f862597a1c"} Dec 10 10:18:33 crc kubenswrapper[4715]: I1210 10:18:33.004426 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:33 crc kubenswrapper[4715]: I1210 10:18:33.005163 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:33 crc kubenswrapper[4715]: I1210 10:18:33.070461 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:33 crc kubenswrapper[4715]: I1210 10:18:33.586374 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:33 crc kubenswrapper[4715]: I1210 10:18:33.647788 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kw89m"] Dec 10 10:18:35 crc kubenswrapper[4715]: I1210 10:18:35.576184 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kw89m" podUID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerName="registry-server" containerID="cri-o://245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348" gracePeriod=2 Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.032263 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.223041 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-catalog-content\") pod \"011917ba-84a4-4cc1-8a22-9e983eac49c1\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.223323 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sszmf\" (UniqueName: \"kubernetes.io/projected/011917ba-84a4-4cc1-8a22-9e983eac49c1-kube-api-access-sszmf\") pod \"011917ba-84a4-4cc1-8a22-9e983eac49c1\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.223415 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-utilities\") pod \"011917ba-84a4-4cc1-8a22-9e983eac49c1\" (UID: \"011917ba-84a4-4cc1-8a22-9e983eac49c1\") " Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.224809 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-utilities" (OuterVolumeSpecName: "utilities") pod "011917ba-84a4-4cc1-8a22-9e983eac49c1" (UID: "011917ba-84a4-4cc1-8a22-9e983eac49c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.228715 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/011917ba-84a4-4cc1-8a22-9e983eac49c1-kube-api-access-sszmf" (OuterVolumeSpecName: "kube-api-access-sszmf") pod "011917ba-84a4-4cc1-8a22-9e983eac49c1" (UID: "011917ba-84a4-4cc1-8a22-9e983eac49c1"). InnerVolumeSpecName "kube-api-access-sszmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.325398 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sszmf\" (UniqueName: \"kubernetes.io/projected/011917ba-84a4-4cc1-8a22-9e983eac49c1-kube-api-access-sszmf\") on node \"crc\" DevicePath \"\"" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.325431 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.592217 4715 generic.go:334] "Generic (PLEG): container finished" podID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerID="245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348" exitCode=0 Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.592264 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kw89m" event={"ID":"011917ba-84a4-4cc1-8a22-9e983eac49c1","Type":"ContainerDied","Data":"245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348"} Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.592303 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kw89m" event={"ID":"011917ba-84a4-4cc1-8a22-9e983eac49c1","Type":"ContainerDied","Data":"00ebffb0e66625cb08ab821c76a2964fb11598111e2e63c4619a8a2ba037b6ca"} Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.592324 4715 scope.go:117] "RemoveContainer" containerID="245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.592383 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kw89m" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.621277 4715 scope.go:117] "RemoveContainer" containerID="46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.647215 4715 scope.go:117] "RemoveContainer" containerID="003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.673205 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "011917ba-84a4-4cc1-8a22-9e983eac49c1" (UID: "011917ba-84a4-4cc1-8a22-9e983eac49c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.693424 4715 scope.go:117] "RemoveContainer" containerID="245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348" Dec 10 10:18:36 crc kubenswrapper[4715]: E1210 10:18:36.694842 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348\": container with ID starting with 245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348 not found: ID does not exist" containerID="245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.694894 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348"} err="failed to get container status \"245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348\": rpc error: code = NotFound desc = could not find container \"245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348\": container with ID starting with 245d0b499192c5078243f0d29072d9abb3eaebcd31a34a6dc182a83da6320348 not found: ID does not exist" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.694946 4715 scope.go:117] "RemoveContainer" containerID="46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40" Dec 10 10:18:36 crc kubenswrapper[4715]: E1210 10:18:36.695577 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40\": container with ID starting with 46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40 not found: ID does not exist" containerID="46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.695628 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40"} err="failed to get container status \"46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40\": rpc error: code = NotFound desc = could not find container \"46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40\": container with ID starting with 46dcd222aa27d078fd82929782df591b6fa963dbfa81a21bf67eca03d9936e40 not found: ID does not exist" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.695644 4715 scope.go:117] "RemoveContainer" containerID="003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494" Dec 10 10:18:36 crc kubenswrapper[4715]: E1210 10:18:36.697082 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494\": container with ID starting with 003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494 not found: ID does not exist" containerID="003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.697157 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494"} err="failed to get container status \"003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494\": rpc error: code = NotFound desc = could not find container \"003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494\": container with ID starting with 003e2509dd52fe57c1b99369c17a58a6c8e7e05a9998afdc5fa9bb85e5667494 not found: ID does not exist" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.732753 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/011917ba-84a4-4cc1-8a22-9e983eac49c1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.943966 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kw89m"] Dec 10 10:18:36 crc kubenswrapper[4715]: I1210 10:18:36.954539 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kw89m"] Dec 10 10:18:37 crc kubenswrapper[4715]: I1210 10:18:37.616124 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="011917ba-84a4-4cc1-8a22-9e983eac49c1" path="/var/lib/kubelet/pods/011917ba-84a4-4cc1-8a22-9e983eac49c1/volumes" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.329904 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 10:18:58 crc kubenswrapper[4715]: E1210 10:18:58.331845 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerName="registry-server" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.331954 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerName="registry-server" Dec 10 10:18:58 crc kubenswrapper[4715]: E1210 10:18:58.332028 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerName="extract-content" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.332091 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerName="extract-content" Dec 10 10:18:58 crc kubenswrapper[4715]: E1210 10:18:58.332165 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerName="extract-utilities" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.332222 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerName="extract-utilities" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.332472 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="011917ba-84a4-4cc1-8a22-9e983eac49c1" containerName="registry-server" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.333271 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.337716 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.337821 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.337952 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-bs8rt" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.339070 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.348329 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.384471 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.384638 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-config-data\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.384709 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.486440 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.486495 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.486538 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.486570 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.486846 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-config-data\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.486944 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.486979 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.487011 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.487041 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zl9g\" (UniqueName: \"kubernetes.io/projected/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-kube-api-access-8zl9g\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.487661 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.488079 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-config-data\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.494580 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.589073 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.589138 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.589170 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.589333 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.589357 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.589381 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zl9g\" (UniqueName: \"kubernetes.io/projected/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-kube-api-access-8zl9g\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.589537 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.590114 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.590160 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.593483 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.594640 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.606581 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zl9g\" (UniqueName: \"kubernetes.io/projected/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-kube-api-access-8zl9g\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.627033 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " pod="openstack/tempest-tests-tempest" Dec 10 10:18:58 crc kubenswrapper[4715]: I1210 10:18:58.658105 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 10:18:59 crc kubenswrapper[4715]: I1210 10:18:59.115554 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 10:18:59 crc kubenswrapper[4715]: I1210 10:18:59.807406 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8","Type":"ContainerStarted","Data":"de3c359ed4d8686e72315a752f2268efdc11dd6d9eb97757f597f03007982ef4"} Dec 10 10:19:32 crc kubenswrapper[4715]: E1210 10:19:32.689462 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 10 10:19:32 crc kubenswrapper[4715]: E1210 10:19:32.690532 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8zl9g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 10:19:32 crc kubenswrapper[4715]: E1210 10:19:32.691803 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" Dec 10 10:19:33 crc kubenswrapper[4715]: E1210 10:19:33.201970 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" Dec 10 10:19:47 crc kubenswrapper[4715]: I1210 10:19:47.341458 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8","Type":"ContainerStarted","Data":"27c08feb4ce7460be21320e5673ca61e9136dc5f287ed5da1685b0150778eea2"} Dec 10 10:19:47 crc kubenswrapper[4715]: I1210 10:19:47.373903 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.386447682 podStartE2EDuration="50.373882534s" podCreationTimestamp="2025-12-10 10:18:57 +0000 UTC" firstStartedPulling="2025-12-10 10:18:59.133394226 +0000 UTC m=+2701.876940467" lastFinishedPulling="2025-12-10 10:19:46.120829058 +0000 UTC m=+2748.864375319" observedRunningTime="2025-12-10 10:19:47.371417314 +0000 UTC m=+2750.114963575" watchObservedRunningTime="2025-12-10 10:19:47.373882534 +0000 UTC m=+2750.117428785" Dec 10 10:20:47 crc kubenswrapper[4715]: I1210 10:20:47.713972 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:20:47 crc kubenswrapper[4715]: I1210 10:20:47.714660 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:21:17 crc kubenswrapper[4715]: I1210 10:21:17.713898 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:21:17 crc kubenswrapper[4715]: I1210 10:21:17.714507 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.300548 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p8jf6"] Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.303821 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.320867 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p8jf6"] Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.361542 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-catalog-content\") pod \"community-operators-p8jf6\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.361728 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbzqk\" (UniqueName: \"kubernetes.io/projected/8f65cc94-c213-406b-b615-a9ad7e86130d-kube-api-access-xbzqk\") pod \"community-operators-p8jf6\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.361960 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-utilities\") pod \"community-operators-p8jf6\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.462655 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-catalog-content\") pod \"community-operators-p8jf6\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.462765 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbzqk\" (UniqueName: \"kubernetes.io/projected/8f65cc94-c213-406b-b615-a9ad7e86130d-kube-api-access-xbzqk\") pod \"community-operators-p8jf6\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.462903 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-utilities\") pod \"community-operators-p8jf6\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.463693 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-catalog-content\") pod \"community-operators-p8jf6\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.463716 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-utilities\") pod \"community-operators-p8jf6\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.492861 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbzqk\" (UniqueName: \"kubernetes.io/projected/8f65cc94-c213-406b-b615-a9ad7e86130d-kube-api-access-xbzqk\") pod \"community-operators-p8jf6\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:18 crc kubenswrapper[4715]: I1210 10:21:18.678312 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:19 crc kubenswrapper[4715]: I1210 10:21:19.382025 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p8jf6"] Dec 10 10:21:19 crc kubenswrapper[4715]: W1210 10:21:19.386279 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f65cc94_c213_406b_b615_a9ad7e86130d.slice/crio-cd523b9bc6ed98a373908016a57baf599042c661c898a3fe1b75f3ee00a4d35c WatchSource:0}: Error finding container cd523b9bc6ed98a373908016a57baf599042c661c898a3fe1b75f3ee00a4d35c: Status 404 returned error can't find the container with id cd523b9bc6ed98a373908016a57baf599042c661c898a3fe1b75f3ee00a4d35c Dec 10 10:21:19 crc kubenswrapper[4715]: I1210 10:21:19.574078 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8jf6" event={"ID":"8f65cc94-c213-406b-b615-a9ad7e86130d","Type":"ContainerStarted","Data":"cd523b9bc6ed98a373908016a57baf599042c661c898a3fe1b75f3ee00a4d35c"} Dec 10 10:21:20 crc kubenswrapper[4715]: I1210 10:21:20.589991 4715 generic.go:334] "Generic (PLEG): container finished" podID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerID="68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72" exitCode=0 Dec 10 10:21:20 crc kubenswrapper[4715]: I1210 10:21:20.590178 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8jf6" event={"ID":"8f65cc94-c213-406b-b615-a9ad7e86130d","Type":"ContainerDied","Data":"68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72"} Dec 10 10:21:20 crc kubenswrapper[4715]: I1210 10:21:20.594544 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 10:21:23 crc kubenswrapper[4715]: I1210 10:21:23.632684 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8jf6" event={"ID":"8f65cc94-c213-406b-b615-a9ad7e86130d","Type":"ContainerStarted","Data":"fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547"} Dec 10 10:21:24 crc kubenswrapper[4715]: I1210 10:21:24.646218 4715 generic.go:334] "Generic (PLEG): container finished" podID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerID="fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547" exitCode=0 Dec 10 10:21:24 crc kubenswrapper[4715]: I1210 10:21:24.646306 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8jf6" event={"ID":"8f65cc94-c213-406b-b615-a9ad7e86130d","Type":"ContainerDied","Data":"fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547"} Dec 10 10:21:26 crc kubenswrapper[4715]: I1210 10:21:26.675999 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8jf6" event={"ID":"8f65cc94-c213-406b-b615-a9ad7e86130d","Type":"ContainerStarted","Data":"8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e"} Dec 10 10:21:26 crc kubenswrapper[4715]: I1210 10:21:26.707165 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p8jf6" podStartSLOduration=3.364416933 podStartE2EDuration="8.707144533s" podCreationTimestamp="2025-12-10 10:21:18 +0000 UTC" firstStartedPulling="2025-12-10 10:21:20.593842482 +0000 UTC m=+2843.337388753" lastFinishedPulling="2025-12-10 10:21:25.936570102 +0000 UTC m=+2848.680116353" observedRunningTime="2025-12-10 10:21:26.699539856 +0000 UTC m=+2849.443086107" watchObservedRunningTime="2025-12-10 10:21:26.707144533 +0000 UTC m=+2849.450690784" Dec 10 10:21:28 crc kubenswrapper[4715]: I1210 10:21:28.680304 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:28 crc kubenswrapper[4715]: I1210 10:21:28.680600 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:28 crc kubenswrapper[4715]: I1210 10:21:28.736627 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:38 crc kubenswrapper[4715]: I1210 10:21:38.752364 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:38 crc kubenswrapper[4715]: I1210 10:21:38.813447 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p8jf6"] Dec 10 10:21:38 crc kubenswrapper[4715]: I1210 10:21:38.813735 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p8jf6" podUID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerName="registry-server" containerID="cri-o://8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e" gracePeriod=2 Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.311708 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.391741 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-catalog-content\") pod \"8f65cc94-c213-406b-b615-a9ad7e86130d\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.391940 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-utilities\") pod \"8f65cc94-c213-406b-b615-a9ad7e86130d\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.392039 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbzqk\" (UniqueName: \"kubernetes.io/projected/8f65cc94-c213-406b-b615-a9ad7e86130d-kube-api-access-xbzqk\") pod \"8f65cc94-c213-406b-b615-a9ad7e86130d\" (UID: \"8f65cc94-c213-406b-b615-a9ad7e86130d\") " Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.392656 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-utilities" (OuterVolumeSpecName: "utilities") pod "8f65cc94-c213-406b-b615-a9ad7e86130d" (UID: "8f65cc94-c213-406b-b615-a9ad7e86130d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.403221 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f65cc94-c213-406b-b615-a9ad7e86130d-kube-api-access-xbzqk" (OuterVolumeSpecName: "kube-api-access-xbzqk") pod "8f65cc94-c213-406b-b615-a9ad7e86130d" (UID: "8f65cc94-c213-406b-b615-a9ad7e86130d"). InnerVolumeSpecName "kube-api-access-xbzqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.451094 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f65cc94-c213-406b-b615-a9ad7e86130d" (UID: "8f65cc94-c213-406b-b615-a9ad7e86130d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.495000 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbzqk\" (UniqueName: \"kubernetes.io/projected/8f65cc94-c213-406b-b615-a9ad7e86130d-kube-api-access-xbzqk\") on node \"crc\" DevicePath \"\"" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.495351 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.495365 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f65cc94-c213-406b-b615-a9ad7e86130d-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.827093 4715 generic.go:334] "Generic (PLEG): container finished" podID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerID="8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e" exitCode=0 Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.827209 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8jf6" event={"ID":"8f65cc94-c213-406b-b615-a9ad7e86130d","Type":"ContainerDied","Data":"8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e"} Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.827804 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p8jf6" event={"ID":"8f65cc94-c213-406b-b615-a9ad7e86130d","Type":"ContainerDied","Data":"cd523b9bc6ed98a373908016a57baf599042c661c898a3fe1b75f3ee00a4d35c"} Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.827871 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p8jf6" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.827867 4715 scope.go:117] "RemoveContainer" containerID="8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.887041 4715 scope.go:117] "RemoveContainer" containerID="fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.894596 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p8jf6"] Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.902967 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p8jf6"] Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.908691 4715 scope.go:117] "RemoveContainer" containerID="68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.965353 4715 scope.go:117] "RemoveContainer" containerID="8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e" Dec 10 10:21:39 crc kubenswrapper[4715]: E1210 10:21:39.965770 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e\": container with ID starting with 8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e not found: ID does not exist" containerID="8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.965833 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e"} err="failed to get container status \"8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e\": rpc error: code = NotFound desc = could not find container \"8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e\": container with ID starting with 8e9ef9f4e501baa9fd9e0c275a55638ff1791044577ac1d55b399dcf259dcb5e not found: ID does not exist" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.965857 4715 scope.go:117] "RemoveContainer" containerID="fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547" Dec 10 10:21:39 crc kubenswrapper[4715]: E1210 10:21:39.966259 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547\": container with ID starting with fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547 not found: ID does not exist" containerID="fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.966391 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547"} err="failed to get container status \"fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547\": rpc error: code = NotFound desc = could not find container \"fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547\": container with ID starting with fa0a23d194a49e9bcc8dbf37e059b84d6c935c5cfaa5fca46e7a77aa005d5547 not found: ID does not exist" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.966467 4715 scope.go:117] "RemoveContainer" containerID="68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72" Dec 10 10:21:39 crc kubenswrapper[4715]: E1210 10:21:39.966826 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72\": container with ID starting with 68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72 not found: ID does not exist" containerID="68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72" Dec 10 10:21:39 crc kubenswrapper[4715]: I1210 10:21:39.966855 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72"} err="failed to get container status \"68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72\": rpc error: code = NotFound desc = could not find container \"68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72\": container with ID starting with 68d57bd39635465e977662f2dd3bc51bad3b31beee3a638261816fab3a8a0d72 not found: ID does not exist" Dec 10 10:21:41 crc kubenswrapper[4715]: I1210 10:21:41.617197 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f65cc94-c213-406b-b615-a9ad7e86130d" path="/var/lib/kubelet/pods/8f65cc94-c213-406b-b615-a9ad7e86130d/volumes" Dec 10 10:21:47 crc kubenswrapper[4715]: I1210 10:21:47.714508 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:21:47 crc kubenswrapper[4715]: I1210 10:21:47.715074 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:21:47 crc kubenswrapper[4715]: I1210 10:21:47.715122 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 10:21:47 crc kubenswrapper[4715]: I1210 10:21:47.902342 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac70cfef39940dc73d054eb086c3ad80da75d6389147830f13b583f862597a1c"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 10:21:47 crc kubenswrapper[4715]: I1210 10:21:47.902435 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://ac70cfef39940dc73d054eb086c3ad80da75d6389147830f13b583f862597a1c" gracePeriod=600 Dec 10 10:21:48 crc kubenswrapper[4715]: I1210 10:21:48.916661 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="ac70cfef39940dc73d054eb086c3ad80da75d6389147830f13b583f862597a1c" exitCode=0 Dec 10 10:21:48 crc kubenswrapper[4715]: I1210 10:21:48.916759 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"ac70cfef39940dc73d054eb086c3ad80da75d6389147830f13b583f862597a1c"} Dec 10 10:21:48 crc kubenswrapper[4715]: I1210 10:21:48.918833 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2"} Dec 10 10:21:48 crc kubenswrapper[4715]: I1210 10:21:48.918964 4715 scope.go:117] "RemoveContainer" containerID="da70a159abf084286f0f44072d5adf98576ea954003eebf2871f894fd1ae934c" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.740371 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2ffpj"] Dec 10 10:23:32 crc kubenswrapper[4715]: E1210 10:23:32.741268 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerName="registry-server" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.741282 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerName="registry-server" Dec 10 10:23:32 crc kubenswrapper[4715]: E1210 10:23:32.741316 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerName="extract-utilities" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.741322 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerName="extract-utilities" Dec 10 10:23:32 crc kubenswrapper[4715]: E1210 10:23:32.741337 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerName="extract-content" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.741345 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerName="extract-content" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.741563 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f65cc94-c213-406b-b615-a9ad7e86130d" containerName="registry-server" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.742978 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.753017 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2ffpj"] Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.831999 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ea9a921-7745-4485-8ba2-b94302d357ad-catalog-content\") pod \"redhat-marketplace-2ffpj\" (UID: \"2ea9a921-7745-4485-8ba2-b94302d357ad\") " pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.832085 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79dqt\" (UniqueName: \"kubernetes.io/projected/2ea9a921-7745-4485-8ba2-b94302d357ad-kube-api-access-79dqt\") pod \"redhat-marketplace-2ffpj\" (UID: \"2ea9a921-7745-4485-8ba2-b94302d357ad\") " pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.832134 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ea9a921-7745-4485-8ba2-b94302d357ad-utilities\") pod \"redhat-marketplace-2ffpj\" (UID: \"2ea9a921-7745-4485-8ba2-b94302d357ad\") " pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.934177 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ea9a921-7745-4485-8ba2-b94302d357ad-catalog-content\") pod \"redhat-marketplace-2ffpj\" (UID: \"2ea9a921-7745-4485-8ba2-b94302d357ad\") " pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.934270 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79dqt\" (UniqueName: \"kubernetes.io/projected/2ea9a921-7745-4485-8ba2-b94302d357ad-kube-api-access-79dqt\") pod \"redhat-marketplace-2ffpj\" (UID: \"2ea9a921-7745-4485-8ba2-b94302d357ad\") " pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.934320 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ea9a921-7745-4485-8ba2-b94302d357ad-utilities\") pod \"redhat-marketplace-2ffpj\" (UID: \"2ea9a921-7745-4485-8ba2-b94302d357ad\") " pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.934693 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ea9a921-7745-4485-8ba2-b94302d357ad-catalog-content\") pod \"redhat-marketplace-2ffpj\" (UID: \"2ea9a921-7745-4485-8ba2-b94302d357ad\") " pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.934753 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ea9a921-7745-4485-8ba2-b94302d357ad-utilities\") pod \"redhat-marketplace-2ffpj\" (UID: \"2ea9a921-7745-4485-8ba2-b94302d357ad\") " pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:32 crc kubenswrapper[4715]: I1210 10:23:32.955695 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79dqt\" (UniqueName: \"kubernetes.io/projected/2ea9a921-7745-4485-8ba2-b94302d357ad-kube-api-access-79dqt\") pod \"redhat-marketplace-2ffpj\" (UID: \"2ea9a921-7745-4485-8ba2-b94302d357ad\") " pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:33 crc kubenswrapper[4715]: I1210 10:23:33.064338 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:33 crc kubenswrapper[4715]: I1210 10:23:33.636048 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2ffpj"] Dec 10 10:23:33 crc kubenswrapper[4715]: I1210 10:23:33.941688 4715 generic.go:334] "Generic (PLEG): container finished" podID="2ea9a921-7745-4485-8ba2-b94302d357ad" containerID="83755235f55b646bc9b39524e68fcd096149c5df9ffb0a126b5c8d72971be109" exitCode=0 Dec 10 10:23:33 crc kubenswrapper[4715]: I1210 10:23:33.941753 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ffpj" event={"ID":"2ea9a921-7745-4485-8ba2-b94302d357ad","Type":"ContainerDied","Data":"83755235f55b646bc9b39524e68fcd096149c5df9ffb0a126b5c8d72971be109"} Dec 10 10:23:33 crc kubenswrapper[4715]: I1210 10:23:33.942017 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ffpj" event={"ID":"2ea9a921-7745-4485-8ba2-b94302d357ad","Type":"ContainerStarted","Data":"acefdff70764f3835110b1d2e5168ba314b18e895bc7ddc614f0d11e6baccfd9"} Dec 10 10:23:37 crc kubenswrapper[4715]: I1210 10:23:37.980258 4715 generic.go:334] "Generic (PLEG): container finished" podID="2ea9a921-7745-4485-8ba2-b94302d357ad" containerID="225c97defe9b68af5b3fb259dc6860bba396b751cdd7666e744350b430fd99de" exitCode=0 Dec 10 10:23:37 crc kubenswrapper[4715]: I1210 10:23:37.980337 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ffpj" event={"ID":"2ea9a921-7745-4485-8ba2-b94302d357ad","Type":"ContainerDied","Data":"225c97defe9b68af5b3fb259dc6860bba396b751cdd7666e744350b430fd99de"} Dec 10 10:23:38 crc kubenswrapper[4715]: I1210 10:23:38.994004 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ffpj" event={"ID":"2ea9a921-7745-4485-8ba2-b94302d357ad","Type":"ContainerStarted","Data":"bb65874765fe6941fa744d6086d9555b71db423cb153c315164c5d69ea25ca19"} Dec 10 10:23:39 crc kubenswrapper[4715]: I1210 10:23:39.014575 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2ffpj" podStartSLOduration=2.476338069 podStartE2EDuration="7.014554138s" podCreationTimestamp="2025-12-10 10:23:32 +0000 UTC" firstStartedPulling="2025-12-10 10:23:33.944313127 +0000 UTC m=+2976.687859378" lastFinishedPulling="2025-12-10 10:23:38.482529196 +0000 UTC m=+2981.226075447" observedRunningTime="2025-12-10 10:23:39.008352681 +0000 UTC m=+2981.751898932" watchObservedRunningTime="2025-12-10 10:23:39.014554138 +0000 UTC m=+2981.758100389" Dec 10 10:23:43 crc kubenswrapper[4715]: I1210 10:23:43.064854 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:43 crc kubenswrapper[4715]: I1210 10:23:43.065556 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:43 crc kubenswrapper[4715]: I1210 10:23:43.113193 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.101423 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2ffpj" Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.170421 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2ffpj"] Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.217345 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sc95c"] Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.217646 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sc95c" podUID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerName="registry-server" containerID="cri-o://7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a" gracePeriod=2 Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.669709 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.864273 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-utilities\") pod \"434fa3ff-6093-4154-82e9-e0612d48cea0\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.864436 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-catalog-content\") pod \"434fa3ff-6093-4154-82e9-e0612d48cea0\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.864707 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbl6w\" (UniqueName: \"kubernetes.io/projected/434fa3ff-6093-4154-82e9-e0612d48cea0-kube-api-access-mbl6w\") pod \"434fa3ff-6093-4154-82e9-e0612d48cea0\" (UID: \"434fa3ff-6093-4154-82e9-e0612d48cea0\") " Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.866296 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-utilities" (OuterVolumeSpecName: "utilities") pod "434fa3ff-6093-4154-82e9-e0612d48cea0" (UID: "434fa3ff-6093-4154-82e9-e0612d48cea0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.877307 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434fa3ff-6093-4154-82e9-e0612d48cea0-kube-api-access-mbl6w" (OuterVolumeSpecName: "kube-api-access-mbl6w") pod "434fa3ff-6093-4154-82e9-e0612d48cea0" (UID: "434fa3ff-6093-4154-82e9-e0612d48cea0"). InnerVolumeSpecName "kube-api-access-mbl6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.897717 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "434fa3ff-6093-4154-82e9-e0612d48cea0" (UID: "434fa3ff-6093-4154-82e9-e0612d48cea0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.978267 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.978329 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbl6w\" (UniqueName: \"kubernetes.io/projected/434fa3ff-6093-4154-82e9-e0612d48cea0-kube-api-access-mbl6w\") on node \"crc\" DevicePath \"\"" Dec 10 10:23:44 crc kubenswrapper[4715]: I1210 10:23:44.978346 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434fa3ff-6093-4154-82e9-e0612d48cea0-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.067907 4715 generic.go:334] "Generic (PLEG): container finished" podID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerID="7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a" exitCode=0 Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.068870 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sc95c" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.071067 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sc95c" event={"ID":"434fa3ff-6093-4154-82e9-e0612d48cea0","Type":"ContainerDied","Data":"7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a"} Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.071120 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sc95c" event={"ID":"434fa3ff-6093-4154-82e9-e0612d48cea0","Type":"ContainerDied","Data":"178ec0d3317b8eff0a855e3ffcf61482f8fea67f39000126d6a809791d063f91"} Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.071138 4715 scope.go:117] "RemoveContainer" containerID="7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.111253 4715 scope.go:117] "RemoveContainer" containerID="9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.117852 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sc95c"] Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.129596 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sc95c"] Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.138000 4715 scope.go:117] "RemoveContainer" containerID="6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.201084 4715 scope.go:117] "RemoveContainer" containerID="7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a" Dec 10 10:23:45 crc kubenswrapper[4715]: E1210 10:23:45.206301 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a\": container with ID starting with 7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a not found: ID does not exist" containerID="7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.206360 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a"} err="failed to get container status \"7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a\": rpc error: code = NotFound desc = could not find container \"7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a\": container with ID starting with 7040a8705a75749587c208dcbabc0738357235053bd5c52c3bb8fd8adfd9c31a not found: ID does not exist" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.206387 4715 scope.go:117] "RemoveContainer" containerID="9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179" Dec 10 10:23:45 crc kubenswrapper[4715]: E1210 10:23:45.207039 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179\": container with ID starting with 9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179 not found: ID does not exist" containerID="9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.207063 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179"} err="failed to get container status \"9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179\": rpc error: code = NotFound desc = could not find container \"9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179\": container with ID starting with 9a48b721ca2b881c31d1cfe36eecda58264a4e2b15cd34f9e0a6ad8719ab1179 not found: ID does not exist" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.207078 4715 scope.go:117] "RemoveContainer" containerID="6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038" Dec 10 10:23:45 crc kubenswrapper[4715]: E1210 10:23:45.207452 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038\": container with ID starting with 6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038 not found: ID does not exist" containerID="6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.207472 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038"} err="failed to get container status \"6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038\": rpc error: code = NotFound desc = could not find container \"6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038\": container with ID starting with 6d09cc35a14952968347bba63932ebf9fb4bc1991978e78aeabd41cb32b12038 not found: ID does not exist" Dec 10 10:23:45 crc kubenswrapper[4715]: I1210 10:23:45.616139 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="434fa3ff-6093-4154-82e9-e0612d48cea0" path="/var/lib/kubelet/pods/434fa3ff-6093-4154-82e9-e0612d48cea0/volumes" Dec 10 10:24:17 crc kubenswrapper[4715]: I1210 10:24:17.714315 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:24:17 crc kubenswrapper[4715]: I1210 10:24:17.714904 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:24:47 crc kubenswrapper[4715]: I1210 10:24:47.714496 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:24:47 crc kubenswrapper[4715]: I1210 10:24:47.714930 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:25:17 crc kubenswrapper[4715]: I1210 10:25:17.714000 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:25:17 crc kubenswrapper[4715]: I1210 10:25:17.714568 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:25:17 crc kubenswrapper[4715]: I1210 10:25:17.714618 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 10:25:17 crc kubenswrapper[4715]: I1210 10:25:17.715395 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 10:25:17 crc kubenswrapper[4715]: I1210 10:25:17.715451 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" gracePeriod=600 Dec 10 10:25:17 crc kubenswrapper[4715]: E1210 10:25:17.837849 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:25:17 crc kubenswrapper[4715]: I1210 10:25:17.961229 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" exitCode=0 Dec 10 10:25:17 crc kubenswrapper[4715]: I1210 10:25:17.961310 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2"} Dec 10 10:25:17 crc kubenswrapper[4715]: I1210 10:25:17.961591 4715 scope.go:117] "RemoveContainer" containerID="ac70cfef39940dc73d054eb086c3ad80da75d6389147830f13b583f862597a1c" Dec 10 10:25:17 crc kubenswrapper[4715]: I1210 10:25:17.962175 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:25:17 crc kubenswrapper[4715]: E1210 10:25:17.962436 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:25:20 crc kubenswrapper[4715]: I1210 10:25:20.929443 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-knnjd"] Dec 10 10:25:20 crc kubenswrapper[4715]: E1210 10:25:20.930765 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerName="extract-content" Dec 10 10:25:20 crc kubenswrapper[4715]: I1210 10:25:20.930805 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerName="extract-content" Dec 10 10:25:20 crc kubenswrapper[4715]: E1210 10:25:20.930861 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerName="extract-utilities" Dec 10 10:25:20 crc kubenswrapper[4715]: I1210 10:25:20.930893 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerName="extract-utilities" Dec 10 10:25:20 crc kubenswrapper[4715]: E1210 10:25:20.930986 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerName="registry-server" Dec 10 10:25:20 crc kubenswrapper[4715]: I1210 10:25:20.931008 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerName="registry-server" Dec 10 10:25:20 crc kubenswrapper[4715]: I1210 10:25:20.931456 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="434fa3ff-6093-4154-82e9-e0612d48cea0" containerName="registry-server" Dec 10 10:25:20 crc kubenswrapper[4715]: I1210 10:25:20.933348 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:20 crc kubenswrapper[4715]: I1210 10:25:20.941119 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-knnjd"] Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.112570 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrn4k\" (UniqueName: \"kubernetes.io/projected/3477bf62-547b-4b46-b952-fa83a9afc465-kube-api-access-qrn4k\") pod \"redhat-operators-knnjd\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.112657 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-catalog-content\") pod \"redhat-operators-knnjd\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.112686 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-utilities\") pod \"redhat-operators-knnjd\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.214865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrn4k\" (UniqueName: \"kubernetes.io/projected/3477bf62-547b-4b46-b952-fa83a9afc465-kube-api-access-qrn4k\") pod \"redhat-operators-knnjd\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.215002 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-catalog-content\") pod \"redhat-operators-knnjd\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.215044 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-utilities\") pod \"redhat-operators-knnjd\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.215586 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-utilities\") pod \"redhat-operators-knnjd\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.215715 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-catalog-content\") pod \"redhat-operators-knnjd\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.236251 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrn4k\" (UniqueName: \"kubernetes.io/projected/3477bf62-547b-4b46-b952-fa83a9afc465-kube-api-access-qrn4k\") pod \"redhat-operators-knnjd\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.271124 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:21 crc kubenswrapper[4715]: I1210 10:25:21.771883 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-knnjd"] Dec 10 10:25:22 crc kubenswrapper[4715]: I1210 10:25:22.004865 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knnjd" event={"ID":"3477bf62-547b-4b46-b952-fa83a9afc465","Type":"ContainerStarted","Data":"ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6"} Dec 10 10:25:22 crc kubenswrapper[4715]: I1210 10:25:22.004965 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knnjd" event={"ID":"3477bf62-547b-4b46-b952-fa83a9afc465","Type":"ContainerStarted","Data":"ccafb28e8ef4f0b5d52ed8e3304c183e245d6d59e15580dd4c08a7848ccd3a25"} Dec 10 10:25:23 crc kubenswrapper[4715]: I1210 10:25:23.020158 4715 generic.go:334] "Generic (PLEG): container finished" podID="3477bf62-547b-4b46-b952-fa83a9afc465" containerID="ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6" exitCode=0 Dec 10 10:25:23 crc kubenswrapper[4715]: I1210 10:25:23.020224 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knnjd" event={"ID":"3477bf62-547b-4b46-b952-fa83a9afc465","Type":"ContainerDied","Data":"ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6"} Dec 10 10:25:24 crc kubenswrapper[4715]: I1210 10:25:24.075656 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knnjd" event={"ID":"3477bf62-547b-4b46-b952-fa83a9afc465","Type":"ContainerStarted","Data":"337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e"} Dec 10 10:25:29 crc kubenswrapper[4715]: I1210 10:25:29.605703 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:25:29 crc kubenswrapper[4715]: E1210 10:25:29.606436 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:25:31 crc kubenswrapper[4715]: I1210 10:25:31.146185 4715 generic.go:334] "Generic (PLEG): container finished" podID="3477bf62-547b-4b46-b952-fa83a9afc465" containerID="337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e" exitCode=0 Dec 10 10:25:31 crc kubenswrapper[4715]: I1210 10:25:31.146231 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knnjd" event={"ID":"3477bf62-547b-4b46-b952-fa83a9afc465","Type":"ContainerDied","Data":"337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e"} Dec 10 10:25:33 crc kubenswrapper[4715]: I1210 10:25:33.178258 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knnjd" event={"ID":"3477bf62-547b-4b46-b952-fa83a9afc465","Type":"ContainerStarted","Data":"00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1"} Dec 10 10:25:33 crc kubenswrapper[4715]: I1210 10:25:33.210376 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-knnjd" podStartSLOduration=3.033692214 podStartE2EDuration="13.210356243s" podCreationTimestamp="2025-12-10 10:25:20 +0000 UTC" firstStartedPulling="2025-12-10 10:25:22.007478278 +0000 UTC m=+3084.751024529" lastFinishedPulling="2025-12-10 10:25:32.184142307 +0000 UTC m=+3094.927688558" observedRunningTime="2025-12-10 10:25:33.200477251 +0000 UTC m=+3095.944023502" watchObservedRunningTime="2025-12-10 10:25:33.210356243 +0000 UTC m=+3095.953902504" Dec 10 10:25:41 crc kubenswrapper[4715]: I1210 10:25:41.271951 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:41 crc kubenswrapper[4715]: I1210 10:25:41.272576 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:42 crc kubenswrapper[4715]: I1210 10:25:42.320267 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-knnjd" podUID="3477bf62-547b-4b46-b952-fa83a9afc465" containerName="registry-server" probeResult="failure" output=< Dec 10 10:25:42 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 10 10:25:42 crc kubenswrapper[4715]: > Dec 10 10:25:42 crc kubenswrapper[4715]: I1210 10:25:42.604990 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:25:42 crc kubenswrapper[4715]: E1210 10:25:42.605383 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:25:51 crc kubenswrapper[4715]: I1210 10:25:51.324667 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:51 crc kubenswrapper[4715]: I1210 10:25:51.376618 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:52 crc kubenswrapper[4715]: I1210 10:25:52.131410 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-knnjd"] Dec 10 10:25:52 crc kubenswrapper[4715]: I1210 10:25:52.356394 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-knnjd" podUID="3477bf62-547b-4b46-b952-fa83a9afc465" containerName="registry-server" containerID="cri-o://00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1" gracePeriod=2 Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.355646 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.388903 4715 generic.go:334] "Generic (PLEG): container finished" podID="3477bf62-547b-4b46-b952-fa83a9afc465" containerID="00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1" exitCode=0 Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.388960 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knnjd" event={"ID":"3477bf62-547b-4b46-b952-fa83a9afc465","Type":"ContainerDied","Data":"00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1"} Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.389030 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knnjd" event={"ID":"3477bf62-547b-4b46-b952-fa83a9afc465","Type":"ContainerDied","Data":"ccafb28e8ef4f0b5d52ed8e3304c183e245d6d59e15580dd4c08a7848ccd3a25"} Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.389058 4715 scope.go:117] "RemoveContainer" containerID="00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.389070 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knnjd" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.408398 4715 scope.go:117] "RemoveContainer" containerID="337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.432014 4715 scope.go:117] "RemoveContainer" containerID="ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.474058 4715 scope.go:117] "RemoveContainer" containerID="00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1" Dec 10 10:25:53 crc kubenswrapper[4715]: E1210 10:25:53.474558 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1\": container with ID starting with 00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1 not found: ID does not exist" containerID="00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.474600 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1"} err="failed to get container status \"00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1\": rpc error: code = NotFound desc = could not find container \"00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1\": container with ID starting with 00e5c4e38ffbeed6a53eee52b17172b77f606eaed8512ff9871ea0fdbe44e0f1 not found: ID does not exist" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.474623 4715 scope.go:117] "RemoveContainer" containerID="337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e" Dec 10 10:25:53 crc kubenswrapper[4715]: E1210 10:25:53.474865 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e\": container with ID starting with 337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e not found: ID does not exist" containerID="337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.474896 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e"} err="failed to get container status \"337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e\": rpc error: code = NotFound desc = could not find container \"337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e\": container with ID starting with 337cbd7586a0829c512534e2a96473a789804aad5c02d0d7117cfdeddba5492e not found: ID does not exist" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.474995 4715 scope.go:117] "RemoveContainer" containerID="ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6" Dec 10 10:25:53 crc kubenswrapper[4715]: E1210 10:25:53.475545 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6\": container with ID starting with ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6 not found: ID does not exist" containerID="ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.475598 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6"} err="failed to get container status \"ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6\": rpc error: code = NotFound desc = could not find container \"ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6\": container with ID starting with ce369a448989346c5ab92abe298056adffc912c45c35df45de6b9d8b3e387ae6 not found: ID does not exist" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.494288 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-catalog-content\") pod \"3477bf62-547b-4b46-b952-fa83a9afc465\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.494399 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrn4k\" (UniqueName: \"kubernetes.io/projected/3477bf62-547b-4b46-b952-fa83a9afc465-kube-api-access-qrn4k\") pod \"3477bf62-547b-4b46-b952-fa83a9afc465\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.494558 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-utilities\") pod \"3477bf62-547b-4b46-b952-fa83a9afc465\" (UID: \"3477bf62-547b-4b46-b952-fa83a9afc465\") " Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.495424 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-utilities" (OuterVolumeSpecName: "utilities") pod "3477bf62-547b-4b46-b952-fa83a9afc465" (UID: "3477bf62-547b-4b46-b952-fa83a9afc465"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.501955 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3477bf62-547b-4b46-b952-fa83a9afc465-kube-api-access-qrn4k" (OuterVolumeSpecName: "kube-api-access-qrn4k") pod "3477bf62-547b-4b46-b952-fa83a9afc465" (UID: "3477bf62-547b-4b46-b952-fa83a9afc465"). InnerVolumeSpecName "kube-api-access-qrn4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.597158 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrn4k\" (UniqueName: \"kubernetes.io/projected/3477bf62-547b-4b46-b952-fa83a9afc465-kube-api-access-qrn4k\") on node \"crc\" DevicePath \"\"" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.597199 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.612121 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3477bf62-547b-4b46-b952-fa83a9afc465" (UID: "3477bf62-547b-4b46-b952-fa83a9afc465"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.698576 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3477bf62-547b-4b46-b952-fa83a9afc465-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.712934 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-knnjd"] Dec 10 10:25:53 crc kubenswrapper[4715]: I1210 10:25:53.722029 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-knnjd"] Dec 10 10:25:55 crc kubenswrapper[4715]: I1210 10:25:55.616369 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3477bf62-547b-4b46-b952-fa83a9afc465" path="/var/lib/kubelet/pods/3477bf62-547b-4b46-b952-fa83a9afc465/volumes" Dec 10 10:25:57 crc kubenswrapper[4715]: I1210 10:25:57.613179 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:25:57 crc kubenswrapper[4715]: E1210 10:25:57.613806 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:26:09 crc kubenswrapper[4715]: I1210 10:26:09.605012 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:26:09 crc kubenswrapper[4715]: E1210 10:26:09.605874 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:26:20 crc kubenswrapper[4715]: I1210 10:26:20.605292 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:26:20 crc kubenswrapper[4715]: E1210 10:26:20.606075 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:26:32 crc kubenswrapper[4715]: I1210 10:26:32.605459 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:26:32 crc kubenswrapper[4715]: E1210 10:26:32.607446 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:26:46 crc kubenswrapper[4715]: I1210 10:26:46.604694 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:26:46 crc kubenswrapper[4715]: E1210 10:26:46.605610 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:26:59 crc kubenswrapper[4715]: I1210 10:26:59.606119 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:26:59 crc kubenswrapper[4715]: E1210 10:26:59.607146 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:27:10 crc kubenswrapper[4715]: I1210 10:27:10.606159 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:27:10 crc kubenswrapper[4715]: E1210 10:27:10.606990 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:27:21 crc kubenswrapper[4715]: I1210 10:27:21.605066 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:27:21 crc kubenswrapper[4715]: E1210 10:27:21.605816 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:27:34 crc kubenswrapper[4715]: I1210 10:27:34.605952 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:27:34 crc kubenswrapper[4715]: E1210 10:27:34.606711 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:27:48 crc kubenswrapper[4715]: I1210 10:27:48.606100 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:27:48 crc kubenswrapper[4715]: E1210 10:27:48.606824 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:28:03 crc kubenswrapper[4715]: I1210 10:28:03.604825 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:28:03 crc kubenswrapper[4715]: E1210 10:28:03.606721 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:28:17 crc kubenswrapper[4715]: I1210 10:28:17.612433 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:28:17 crc kubenswrapper[4715]: E1210 10:28:17.613259 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.081428 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6lhbn"] Dec 10 10:28:31 crc kubenswrapper[4715]: E1210 10:28:31.082522 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3477bf62-547b-4b46-b952-fa83a9afc465" containerName="extract-utilities" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.082552 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3477bf62-547b-4b46-b952-fa83a9afc465" containerName="extract-utilities" Dec 10 10:28:31 crc kubenswrapper[4715]: E1210 10:28:31.082574 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3477bf62-547b-4b46-b952-fa83a9afc465" containerName="extract-content" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.082582 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3477bf62-547b-4b46-b952-fa83a9afc465" containerName="extract-content" Dec 10 10:28:31 crc kubenswrapper[4715]: E1210 10:28:31.082617 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3477bf62-547b-4b46-b952-fa83a9afc465" containerName="registry-server" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.082624 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3477bf62-547b-4b46-b952-fa83a9afc465" containerName="registry-server" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.082933 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3477bf62-547b-4b46-b952-fa83a9afc465" containerName="registry-server" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.084591 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.094582 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6lhbn"] Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.213783 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jnqc\" (UniqueName: \"kubernetes.io/projected/bc172f3d-d7ab-4675-8720-107c38a9f90a-kube-api-access-2jnqc\") pod \"certified-operators-6lhbn\" (UID: \"bc172f3d-d7ab-4675-8720-107c38a9f90a\") " pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.213837 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc172f3d-d7ab-4675-8720-107c38a9f90a-utilities\") pod \"certified-operators-6lhbn\" (UID: \"bc172f3d-d7ab-4675-8720-107c38a9f90a\") " pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.214184 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc172f3d-d7ab-4675-8720-107c38a9f90a-catalog-content\") pod \"certified-operators-6lhbn\" (UID: \"bc172f3d-d7ab-4675-8720-107c38a9f90a\") " pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.315769 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jnqc\" (UniqueName: \"kubernetes.io/projected/bc172f3d-d7ab-4675-8720-107c38a9f90a-kube-api-access-2jnqc\") pod \"certified-operators-6lhbn\" (UID: \"bc172f3d-d7ab-4675-8720-107c38a9f90a\") " pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.315826 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc172f3d-d7ab-4675-8720-107c38a9f90a-utilities\") pod \"certified-operators-6lhbn\" (UID: \"bc172f3d-d7ab-4675-8720-107c38a9f90a\") " pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.315976 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc172f3d-d7ab-4675-8720-107c38a9f90a-catalog-content\") pod \"certified-operators-6lhbn\" (UID: \"bc172f3d-d7ab-4675-8720-107c38a9f90a\") " pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.316410 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc172f3d-d7ab-4675-8720-107c38a9f90a-utilities\") pod \"certified-operators-6lhbn\" (UID: \"bc172f3d-d7ab-4675-8720-107c38a9f90a\") " pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.316514 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc172f3d-d7ab-4675-8720-107c38a9f90a-catalog-content\") pod \"certified-operators-6lhbn\" (UID: \"bc172f3d-d7ab-4675-8720-107c38a9f90a\") " pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.357694 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jnqc\" (UniqueName: \"kubernetes.io/projected/bc172f3d-d7ab-4675-8720-107c38a9f90a-kube-api-access-2jnqc\") pod \"certified-operators-6lhbn\" (UID: \"bc172f3d-d7ab-4675-8720-107c38a9f90a\") " pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.416488 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:31 crc kubenswrapper[4715]: I1210 10:28:31.606603 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:28:31 crc kubenswrapper[4715]: E1210 10:28:31.607373 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:28:32 crc kubenswrapper[4715]: I1210 10:28:32.011627 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6lhbn"] Dec 10 10:28:32 crc kubenswrapper[4715]: I1210 10:28:32.894512 4715 generic.go:334] "Generic (PLEG): container finished" podID="bc172f3d-d7ab-4675-8720-107c38a9f90a" containerID="de91d31edbfe31df4a760f2df0b15c6d8461ad17711f24cc4b4243d6f822af1c" exitCode=0 Dec 10 10:28:32 crc kubenswrapper[4715]: I1210 10:28:32.894635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lhbn" event={"ID":"bc172f3d-d7ab-4675-8720-107c38a9f90a","Type":"ContainerDied","Data":"de91d31edbfe31df4a760f2df0b15c6d8461ad17711f24cc4b4243d6f822af1c"} Dec 10 10:28:32 crc kubenswrapper[4715]: I1210 10:28:32.894823 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lhbn" event={"ID":"bc172f3d-d7ab-4675-8720-107c38a9f90a","Type":"ContainerStarted","Data":"4672554c888303f19b004c095992ff95e03646510435e0d1278681b27c359892"} Dec 10 10:28:32 crc kubenswrapper[4715]: I1210 10:28:32.898019 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 10:28:39 crc kubenswrapper[4715]: I1210 10:28:39.965725 4715 generic.go:334] "Generic (PLEG): container finished" podID="bc172f3d-d7ab-4675-8720-107c38a9f90a" containerID="dfce31723b9021688f4cd205e08d65b66fecdd7c35f7b17b1ad768d326289b71" exitCode=0 Dec 10 10:28:39 crc kubenswrapper[4715]: I1210 10:28:39.965881 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lhbn" event={"ID":"bc172f3d-d7ab-4675-8720-107c38a9f90a","Type":"ContainerDied","Data":"dfce31723b9021688f4cd205e08d65b66fecdd7c35f7b17b1ad768d326289b71"} Dec 10 10:28:40 crc kubenswrapper[4715]: I1210 10:28:40.978691 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lhbn" event={"ID":"bc172f3d-d7ab-4675-8720-107c38a9f90a","Type":"ContainerStarted","Data":"2fe2833660893fddefbbee69f32b5beaa9ff67668636f18e5c25f2cd02504b5b"} Dec 10 10:28:40 crc kubenswrapper[4715]: I1210 10:28:40.997293 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6lhbn" podStartSLOduration=2.347705019 podStartE2EDuration="9.99727545s" podCreationTimestamp="2025-12-10 10:28:31 +0000 UTC" firstStartedPulling="2025-12-10 10:28:32.897694629 +0000 UTC m=+3275.641240880" lastFinishedPulling="2025-12-10 10:28:40.54726506 +0000 UTC m=+3283.290811311" observedRunningTime="2025-12-10 10:28:40.994232323 +0000 UTC m=+3283.737778574" watchObservedRunningTime="2025-12-10 10:28:40.99727545 +0000 UTC m=+3283.740821711" Dec 10 10:28:41 crc kubenswrapper[4715]: I1210 10:28:41.416846 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:41 crc kubenswrapper[4715]: I1210 10:28:41.417222 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:42 crc kubenswrapper[4715]: I1210 10:28:42.467581 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-6lhbn" podUID="bc172f3d-d7ab-4675-8720-107c38a9f90a" containerName="registry-server" probeResult="failure" output=< Dec 10 10:28:42 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 10 10:28:42 crc kubenswrapper[4715]: > Dec 10 10:28:42 crc kubenswrapper[4715]: I1210 10:28:42.605323 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:28:42 crc kubenswrapper[4715]: E1210 10:28:42.605608 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:28:51 crc kubenswrapper[4715]: I1210 10:28:51.468652 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:51 crc kubenswrapper[4715]: I1210 10:28:51.529470 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6lhbn" Dec 10 10:28:51 crc kubenswrapper[4715]: I1210 10:28:51.599164 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6lhbn"] Dec 10 10:28:51 crc kubenswrapper[4715]: I1210 10:28:51.712728 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tjntt"] Dec 10 10:28:51 crc kubenswrapper[4715]: I1210 10:28:51.712984 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tjntt" podUID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerName="registry-server" containerID="cri-o://9d6415e023064b246fc7943f6671a6d94a80084d3abaa00c021a524954530a79" gracePeriod=2 Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.094730 4715 generic.go:334] "Generic (PLEG): container finished" podID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerID="9d6415e023064b246fc7943f6671a6d94a80084d3abaa00c021a524954530a79" exitCode=0 Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.094931 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjntt" event={"ID":"4e744f3a-62da-4f4e-a7ae-5fec5c47873f","Type":"ContainerDied","Data":"9d6415e023064b246fc7943f6671a6d94a80084d3abaa00c021a524954530a79"} Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.266603 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjntt" Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.400285 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-catalog-content\") pod \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.400801 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn5wp\" (UniqueName: \"kubernetes.io/projected/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-kube-api-access-bn5wp\") pod \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.400930 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-utilities\") pod \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\" (UID: \"4e744f3a-62da-4f4e-a7ae-5fec5c47873f\") " Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.402840 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-utilities" (OuterVolumeSpecName: "utilities") pod "4e744f3a-62da-4f4e-a7ae-5fec5c47873f" (UID: "4e744f3a-62da-4f4e-a7ae-5fec5c47873f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.423613 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-kube-api-access-bn5wp" (OuterVolumeSpecName: "kube-api-access-bn5wp") pod "4e744f3a-62da-4f4e-a7ae-5fec5c47873f" (UID: "4e744f3a-62da-4f4e-a7ae-5fec5c47873f"). InnerVolumeSpecName "kube-api-access-bn5wp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.471886 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e744f3a-62da-4f4e-a7ae-5fec5c47873f" (UID: "4e744f3a-62da-4f4e-a7ae-5fec5c47873f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.504034 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.504082 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn5wp\" (UniqueName: \"kubernetes.io/projected/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-kube-api-access-bn5wp\") on node \"crc\" DevicePath \"\"" Dec 10 10:28:52 crc kubenswrapper[4715]: I1210 10:28:52.504095 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e744f3a-62da-4f4e-a7ae-5fec5c47873f-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:28:53 crc kubenswrapper[4715]: I1210 10:28:53.121604 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tjntt" event={"ID":"4e744f3a-62da-4f4e-a7ae-5fec5c47873f","Type":"ContainerDied","Data":"9e7ef8efe9e13ea8d9e027e850dd280c759b228eed467de182afb4649906bccb"} Dec 10 10:28:53 crc kubenswrapper[4715]: I1210 10:28:53.121653 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tjntt" Dec 10 10:28:53 crc kubenswrapper[4715]: I1210 10:28:53.121666 4715 scope.go:117] "RemoveContainer" containerID="9d6415e023064b246fc7943f6671a6d94a80084d3abaa00c021a524954530a79" Dec 10 10:28:53 crc kubenswrapper[4715]: I1210 10:28:53.159495 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tjntt"] Dec 10 10:28:53 crc kubenswrapper[4715]: I1210 10:28:53.160673 4715 scope.go:117] "RemoveContainer" containerID="f47200b14f28e4f2e7233a4d688bab7094c18c62ae3b1856ebb1c3b3ac8c1254" Dec 10 10:28:53 crc kubenswrapper[4715]: I1210 10:28:53.173436 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tjntt"] Dec 10 10:28:53 crc kubenswrapper[4715]: I1210 10:28:53.194305 4715 scope.go:117] "RemoveContainer" containerID="7a5e1aa3f73ff0a3a8cdba61fce3c35b91d6dd0715667eeacab6bacfdb3a1fcd" Dec 10 10:28:53 crc kubenswrapper[4715]: I1210 10:28:53.618175 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" path="/var/lib/kubelet/pods/4e744f3a-62da-4f4e-a7ae-5fec5c47873f/volumes" Dec 10 10:28:56 crc kubenswrapper[4715]: I1210 10:28:56.605694 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:28:56 crc kubenswrapper[4715]: E1210 10:28:56.606359 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:29:10 crc kubenswrapper[4715]: I1210 10:29:10.606029 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:29:10 crc kubenswrapper[4715]: E1210 10:29:10.606890 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:29:21 crc kubenswrapper[4715]: I1210 10:29:21.605525 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:29:21 crc kubenswrapper[4715]: E1210 10:29:21.606445 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:29:33 crc kubenswrapper[4715]: I1210 10:29:33.605992 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:29:33 crc kubenswrapper[4715]: E1210 10:29:33.607010 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:29:44 crc kubenswrapper[4715]: I1210 10:29:44.604973 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:29:44 crc kubenswrapper[4715]: E1210 10:29:44.605553 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:29:58 crc kubenswrapper[4715]: I1210 10:29:58.605448 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:29:58 crc kubenswrapper[4715]: E1210 10:29:58.607353 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.156525 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh"] Dec 10 10:30:00 crc kubenswrapper[4715]: E1210 10:30:00.157496 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerName="extract-utilities" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.157516 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerName="extract-utilities" Dec 10 10:30:00 crc kubenswrapper[4715]: E1210 10:30:00.157567 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerName="registry-server" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.157575 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerName="registry-server" Dec 10 10:30:00 crc kubenswrapper[4715]: E1210 10:30:00.157598 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerName="extract-content" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.157603 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerName="extract-content" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.157819 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e744f3a-62da-4f4e-a7ae-5fec5c47873f" containerName="registry-server" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.158662 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.162438 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.162577 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.168770 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh"] Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.292802 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jb7f\" (UniqueName: \"kubernetes.io/projected/153c868c-1b06-4a21-bc41-6362feb1bb2a-kube-api-access-6jb7f\") pod \"collect-profiles-29422710-4tjnh\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.292946 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/153c868c-1b06-4a21-bc41-6362feb1bb2a-secret-volume\") pod \"collect-profiles-29422710-4tjnh\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.293019 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/153c868c-1b06-4a21-bc41-6362feb1bb2a-config-volume\") pod \"collect-profiles-29422710-4tjnh\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.395129 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/153c868c-1b06-4a21-bc41-6362feb1bb2a-secret-volume\") pod \"collect-profiles-29422710-4tjnh\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.395292 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/153c868c-1b06-4a21-bc41-6362feb1bb2a-config-volume\") pod \"collect-profiles-29422710-4tjnh\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.395391 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jb7f\" (UniqueName: \"kubernetes.io/projected/153c868c-1b06-4a21-bc41-6362feb1bb2a-kube-api-access-6jb7f\") pod \"collect-profiles-29422710-4tjnh\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.396322 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/153c868c-1b06-4a21-bc41-6362feb1bb2a-config-volume\") pod \"collect-profiles-29422710-4tjnh\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.405345 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/153c868c-1b06-4a21-bc41-6362feb1bb2a-secret-volume\") pod \"collect-profiles-29422710-4tjnh\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.414094 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jb7f\" (UniqueName: \"kubernetes.io/projected/153c868c-1b06-4a21-bc41-6362feb1bb2a-kube-api-access-6jb7f\") pod \"collect-profiles-29422710-4tjnh\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.483082 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:00 crc kubenswrapper[4715]: I1210 10:30:00.935503 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh"] Dec 10 10:30:01 crc kubenswrapper[4715]: I1210 10:30:01.771850 4715 generic.go:334] "Generic (PLEG): container finished" podID="153c868c-1b06-4a21-bc41-6362feb1bb2a" containerID="e44f3116ad7617303b2dec3b053da3fb985d47ed4a607f091cb7de0d16eb061b" exitCode=0 Dec 10 10:30:01 crc kubenswrapper[4715]: I1210 10:30:01.771951 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" event={"ID":"153c868c-1b06-4a21-bc41-6362feb1bb2a","Type":"ContainerDied","Data":"e44f3116ad7617303b2dec3b053da3fb985d47ed4a607f091cb7de0d16eb061b"} Dec 10 10:30:01 crc kubenswrapper[4715]: I1210 10:30:01.772444 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" event={"ID":"153c868c-1b06-4a21-bc41-6362feb1bb2a","Type":"ContainerStarted","Data":"59c57a11d719b72d16bfff5a1d7e722fb4d29f12b031e068886854ac7069e2dd"} Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.177400 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.356052 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/153c868c-1b06-4a21-bc41-6362feb1bb2a-config-volume\") pod \"153c868c-1b06-4a21-bc41-6362feb1bb2a\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.356266 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/153c868c-1b06-4a21-bc41-6362feb1bb2a-secret-volume\") pod \"153c868c-1b06-4a21-bc41-6362feb1bb2a\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.356402 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jb7f\" (UniqueName: \"kubernetes.io/projected/153c868c-1b06-4a21-bc41-6362feb1bb2a-kube-api-access-6jb7f\") pod \"153c868c-1b06-4a21-bc41-6362feb1bb2a\" (UID: \"153c868c-1b06-4a21-bc41-6362feb1bb2a\") " Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.357328 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/153c868c-1b06-4a21-bc41-6362feb1bb2a-config-volume" (OuterVolumeSpecName: "config-volume") pod "153c868c-1b06-4a21-bc41-6362feb1bb2a" (UID: "153c868c-1b06-4a21-bc41-6362feb1bb2a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.362415 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/153c868c-1b06-4a21-bc41-6362feb1bb2a-kube-api-access-6jb7f" (OuterVolumeSpecName: "kube-api-access-6jb7f") pod "153c868c-1b06-4a21-bc41-6362feb1bb2a" (UID: "153c868c-1b06-4a21-bc41-6362feb1bb2a"). InnerVolumeSpecName "kube-api-access-6jb7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.365066 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/153c868c-1b06-4a21-bc41-6362feb1bb2a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "153c868c-1b06-4a21-bc41-6362feb1bb2a" (UID: "153c868c-1b06-4a21-bc41-6362feb1bb2a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.460165 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/153c868c-1b06-4a21-bc41-6362feb1bb2a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.460238 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/153c868c-1b06-4a21-bc41-6362feb1bb2a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.460265 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jb7f\" (UniqueName: \"kubernetes.io/projected/153c868c-1b06-4a21-bc41-6362feb1bb2a-kube-api-access-6jb7f\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.794946 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" event={"ID":"153c868c-1b06-4a21-bc41-6362feb1bb2a","Type":"ContainerDied","Data":"59c57a11d719b72d16bfff5a1d7e722fb4d29f12b031e068886854ac7069e2dd"} Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.795298 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59c57a11d719b72d16bfff5a1d7e722fb4d29f12b031e068886854ac7069e2dd" Dec 10 10:30:03 crc kubenswrapper[4715]: I1210 10:30:03.794992 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422710-4tjnh" Dec 10 10:30:04 crc kubenswrapper[4715]: I1210 10:30:04.261805 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq"] Dec 10 10:30:04 crc kubenswrapper[4715]: I1210 10:30:04.270360 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422665-j4qmq"] Dec 10 10:30:05 crc kubenswrapper[4715]: I1210 10:30:05.620312 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e404a964-ca8d-49f5-a026-eeea3c71582c" path="/var/lib/kubelet/pods/e404a964-ca8d-49f5-a026-eeea3c71582c/volumes" Dec 10 10:30:09 crc kubenswrapper[4715]: I1210 10:30:09.605444 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:30:09 crc kubenswrapper[4715]: E1210 10:30:09.606313 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:30:22 crc kubenswrapper[4715]: I1210 10:30:22.605132 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:30:22 crc kubenswrapper[4715]: I1210 10:30:22.998585 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"62c20bae2d69f89d01d1894e46f3b38c45cea4c280173ffb7123e37e92db2bf8"} Dec 10 10:30:32 crc kubenswrapper[4715]: I1210 10:30:32.864804 4715 scope.go:117] "RemoveContainer" containerID="b87daa8256509516b33fbe9f34846c8af055f3366fd79998bbd45f5a0fa2a5ed" Dec 10 10:30:51 crc kubenswrapper[4715]: I1210 10:30:51.282607 4715 generic.go:334] "Generic (PLEG): container finished" podID="a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" containerID="27c08feb4ce7460be21320e5673ca61e9136dc5f287ed5da1685b0150778eea2" exitCode=0 Dec 10 10:30:51 crc kubenswrapper[4715]: I1210 10:30:51.282722 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8","Type":"ContainerDied","Data":"27c08feb4ce7460be21320e5673ca61e9136dc5f287ed5da1685b0150778eea2"} Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.716037 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.825584 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ssh-key\") pod \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.825682 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config-secret\") pod \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.825706 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-workdir\") pod \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.825732 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zl9g\" (UniqueName: \"kubernetes.io/projected/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-kube-api-access-8zl9g\") pod \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.827879 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ca-certs\") pod \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.828204 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-temporary\") pod \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.828281 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.828337 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config\") pod \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.828363 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-config-data\") pod \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\" (UID: \"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8\") " Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.829303 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" (UID: "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.830225 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-config-data" (OuterVolumeSpecName: "config-data") pod "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" (UID: "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.831156 4715 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.833878 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-kube-api-access-8zl9g" (OuterVolumeSpecName: "kube-api-access-8zl9g") pod "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" (UID: "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8"). InnerVolumeSpecName "kube-api-access-8zl9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.835151 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" (UID: "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.841569 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" (UID: "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.864700 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" (UID: "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.865372 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" (UID: "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.867575 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" (UID: "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.892735 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" (UID: "a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.933195 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.933252 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.933285 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.933295 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.933308 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.933317 4715 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.933330 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zl9g\" (UniqueName: \"kubernetes.io/projected/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-kube-api-access-8zl9g\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.933339 4715 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:52 crc kubenswrapper[4715]: I1210 10:30:52.958282 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 10 10:30:53 crc kubenswrapper[4715]: I1210 10:30:53.035785 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 10 10:30:53 crc kubenswrapper[4715]: I1210 10:30:53.306308 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8","Type":"ContainerDied","Data":"de3c359ed4d8686e72315a752f2268efdc11dd6d9eb97757f597f03007982ef4"} Dec 10 10:30:53 crc kubenswrapper[4715]: I1210 10:30:53.306363 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de3c359ed4d8686e72315a752f2268efdc11dd6d9eb97757f597f03007982ef4" Dec 10 10:30:53 crc kubenswrapper[4715]: I1210 10:30:53.306424 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.354000 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 10:30:59 crc kubenswrapper[4715]: E1210 10:30:59.355336 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" containerName="tempest-tests-tempest-tests-runner" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.355355 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" containerName="tempest-tests-tempest-tests-runner" Dec 10 10:30:59 crc kubenswrapper[4715]: E1210 10:30:59.355392 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="153c868c-1b06-4a21-bc41-6362feb1bb2a" containerName="collect-profiles" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.355400 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="153c868c-1b06-4a21-bc41-6362feb1bb2a" containerName="collect-profiles" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.355782 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8" containerName="tempest-tests-tempest-tests-runner" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.355806 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="153c868c-1b06-4a21-bc41-6362feb1bb2a" containerName="collect-profiles" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.356686 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.360513 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-bs8rt" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.373651 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.478101 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59mvn\" (UniqueName: \"kubernetes.io/projected/40f62965-1d20-4eba-803a-b513c07e489f-kube-api-access-59mvn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"40f62965-1d20-4eba-803a-b513c07e489f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.478458 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"40f62965-1d20-4eba-803a-b513c07e489f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.580614 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59mvn\" (UniqueName: \"kubernetes.io/projected/40f62965-1d20-4eba-803a-b513c07e489f-kube-api-access-59mvn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"40f62965-1d20-4eba-803a-b513c07e489f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.580816 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"40f62965-1d20-4eba-803a-b513c07e489f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.581413 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"40f62965-1d20-4eba-803a-b513c07e489f\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.608819 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59mvn\" (UniqueName: \"kubernetes.io/projected/40f62965-1d20-4eba-803a-b513c07e489f-kube-api-access-59mvn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"40f62965-1d20-4eba-803a-b513c07e489f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.619543 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"40f62965-1d20-4eba-803a-b513c07e489f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 10:30:59 crc kubenswrapper[4715]: I1210 10:30:59.689148 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 10:31:00 crc kubenswrapper[4715]: I1210 10:31:00.248953 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 10:31:00 crc kubenswrapper[4715]: I1210 10:31:00.385291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"40f62965-1d20-4eba-803a-b513c07e489f","Type":"ContainerStarted","Data":"680a5b9fe9fd04fca385a777ae185e4f7327b82ec828b59bc9c40a891d75a182"} Dec 10 10:31:02 crc kubenswrapper[4715]: I1210 10:31:02.405291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"40f62965-1d20-4eba-803a-b513c07e489f","Type":"ContainerStarted","Data":"39a95ca3c49e6dd79fd15d576a5561cd96f9cf9cbd160141c34beea95e2df21e"} Dec 10 10:31:02 crc kubenswrapper[4715]: I1210 10:31:02.421645 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.448796282 podStartE2EDuration="3.421625783s" podCreationTimestamp="2025-12-10 10:30:59 +0000 UTC" firstStartedPulling="2025-12-10 10:31:00.263413317 +0000 UTC m=+3423.006959568" lastFinishedPulling="2025-12-10 10:31:01.236242818 +0000 UTC m=+3423.979789069" observedRunningTime="2025-12-10 10:31:02.416211068 +0000 UTC m=+3425.159757329" watchObservedRunningTime="2025-12-10 10:31:02.421625783 +0000 UTC m=+3425.165172034" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.099901 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-k97pd/must-gather-2j6sm"] Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.102694 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/must-gather-2j6sm" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.104540 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-k97pd"/"default-dockercfg-cwdph" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.104563 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-k97pd"/"kube-root-ca.crt" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.104643 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-k97pd"/"openshift-service-ca.crt" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.118642 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-k97pd/must-gather-2j6sm"] Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.223488 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6c19a077-d0e9-46ab-8c3b-ab9406292b85-must-gather-output\") pod \"must-gather-2j6sm\" (UID: \"6c19a077-d0e9-46ab-8c3b-ab9406292b85\") " pod="openshift-must-gather-k97pd/must-gather-2j6sm" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.223608 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9fsn\" (UniqueName: \"kubernetes.io/projected/6c19a077-d0e9-46ab-8c3b-ab9406292b85-kube-api-access-x9fsn\") pod \"must-gather-2j6sm\" (UID: \"6c19a077-d0e9-46ab-8c3b-ab9406292b85\") " pod="openshift-must-gather-k97pd/must-gather-2j6sm" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.325787 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9fsn\" (UniqueName: \"kubernetes.io/projected/6c19a077-d0e9-46ab-8c3b-ab9406292b85-kube-api-access-x9fsn\") pod \"must-gather-2j6sm\" (UID: \"6c19a077-d0e9-46ab-8c3b-ab9406292b85\") " pod="openshift-must-gather-k97pd/must-gather-2j6sm" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.325955 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6c19a077-d0e9-46ab-8c3b-ab9406292b85-must-gather-output\") pod \"must-gather-2j6sm\" (UID: \"6c19a077-d0e9-46ab-8c3b-ab9406292b85\") " pod="openshift-must-gather-k97pd/must-gather-2j6sm" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.326536 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6c19a077-d0e9-46ab-8c3b-ab9406292b85-must-gather-output\") pod \"must-gather-2j6sm\" (UID: \"6c19a077-d0e9-46ab-8c3b-ab9406292b85\") " pod="openshift-must-gather-k97pd/must-gather-2j6sm" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.344595 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9fsn\" (UniqueName: \"kubernetes.io/projected/6c19a077-d0e9-46ab-8c3b-ab9406292b85-kube-api-access-x9fsn\") pod \"must-gather-2j6sm\" (UID: \"6c19a077-d0e9-46ab-8c3b-ab9406292b85\") " pod="openshift-must-gather-k97pd/must-gather-2j6sm" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.430875 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/must-gather-2j6sm" Dec 10 10:31:24 crc kubenswrapper[4715]: I1210 10:31:24.884225 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-k97pd/must-gather-2j6sm"] Dec 10 10:31:24 crc kubenswrapper[4715]: W1210 10:31:24.885103 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c19a077_d0e9_46ab_8c3b_ab9406292b85.slice/crio-d38bff8dcb0f3833abaf6ef2ecf77e044d007de1b698f30762965fe0109ab68f WatchSource:0}: Error finding container d38bff8dcb0f3833abaf6ef2ecf77e044d007de1b698f30762965fe0109ab68f: Status 404 returned error can't find the container with id d38bff8dcb0f3833abaf6ef2ecf77e044d007de1b698f30762965fe0109ab68f Dec 10 10:31:25 crc kubenswrapper[4715]: I1210 10:31:25.690106 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/must-gather-2j6sm" event={"ID":"6c19a077-d0e9-46ab-8c3b-ab9406292b85","Type":"ContainerStarted","Data":"d38bff8dcb0f3833abaf6ef2ecf77e044d007de1b698f30762965fe0109ab68f"} Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.065290 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hrt7p"] Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.069047 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.102654 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hrt7p"] Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.206405 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-catalog-content\") pod \"community-operators-hrt7p\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.206562 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-utilities\") pod \"community-operators-hrt7p\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.206600 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44p5b\" (UniqueName: \"kubernetes.io/projected/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-kube-api-access-44p5b\") pod \"community-operators-hrt7p\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.309038 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-utilities\") pod \"community-operators-hrt7p\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.309189 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44p5b\" (UniqueName: \"kubernetes.io/projected/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-kube-api-access-44p5b\") pod \"community-operators-hrt7p\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.309281 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-catalog-content\") pod \"community-operators-hrt7p\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.310146 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-catalog-content\") pod \"community-operators-hrt7p\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.310554 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-utilities\") pod \"community-operators-hrt7p\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.336944 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44p5b\" (UniqueName: \"kubernetes.io/projected/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-kube-api-access-44p5b\") pod \"community-operators-hrt7p\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:28 crc kubenswrapper[4715]: I1210 10:31:28.420283 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:33 crc kubenswrapper[4715]: I1210 10:31:33.881158 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hrt7p"] Dec 10 10:31:34 crc kubenswrapper[4715]: I1210 10:31:34.808705 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/must-gather-2j6sm" event={"ID":"6c19a077-d0e9-46ab-8c3b-ab9406292b85","Type":"ContainerStarted","Data":"fcd7ff9db8716ad243c700648e501f79292c69e793f8b9df09334f8bee319e9d"} Dec 10 10:31:34 crc kubenswrapper[4715]: I1210 10:31:34.809432 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/must-gather-2j6sm" event={"ID":"6c19a077-d0e9-46ab-8c3b-ab9406292b85","Type":"ContainerStarted","Data":"6ff4a052eab48f4016cf6b4e1a1cfcc20f9c15df35234d31e09c297b37fc2f2d"} Dec 10 10:31:34 crc kubenswrapper[4715]: I1210 10:31:34.813195 4715 generic.go:334] "Generic (PLEG): container finished" podID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerID="61b4020ed5f339956ed930c03c4ca46993dcc0496f5c36c2bd498a2c8116d5ab" exitCode=0 Dec 10 10:31:34 crc kubenswrapper[4715]: I1210 10:31:34.813248 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrt7p" event={"ID":"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c","Type":"ContainerDied","Data":"61b4020ed5f339956ed930c03c4ca46993dcc0496f5c36c2bd498a2c8116d5ab"} Dec 10 10:31:34 crc kubenswrapper[4715]: I1210 10:31:34.813280 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrt7p" event={"ID":"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c","Type":"ContainerStarted","Data":"79f2dc56d4b67294e059d6a953b8b56f8f58f5cf1940d9334b2e88f48851282b"} Dec 10 10:31:34 crc kubenswrapper[4715]: I1210 10:31:34.834560 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-k97pd/must-gather-2j6sm" podStartSLOduration=2.231705183 podStartE2EDuration="10.834540936s" podCreationTimestamp="2025-12-10 10:31:24 +0000 UTC" firstStartedPulling="2025-12-10 10:31:24.888099966 +0000 UTC m=+3447.631646217" lastFinishedPulling="2025-12-10 10:31:33.490935709 +0000 UTC m=+3456.234481970" observedRunningTime="2025-12-10 10:31:34.832288151 +0000 UTC m=+3457.575834412" watchObservedRunningTime="2025-12-10 10:31:34.834540936 +0000 UTC m=+3457.578087187" Dec 10 10:31:36 crc kubenswrapper[4715]: I1210 10:31:36.835040 4715 generic.go:334] "Generic (PLEG): container finished" podID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerID="074ac3ad85de30771334e668f7948bec8e97f5cb4f7f42081052d44fc22ee4da" exitCode=0 Dec 10 10:31:36 crc kubenswrapper[4715]: I1210 10:31:36.835557 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrt7p" event={"ID":"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c","Type":"ContainerDied","Data":"074ac3ad85de30771334e668f7948bec8e97f5cb4f7f42081052d44fc22ee4da"} Dec 10 10:31:38 crc kubenswrapper[4715]: I1210 10:31:38.093166 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-k97pd/crc-debug-ksgws"] Dec 10 10:31:38 crc kubenswrapper[4715]: I1210 10:31:38.094938 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-ksgws" Dec 10 10:31:38 crc kubenswrapper[4715]: I1210 10:31:38.169350 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbc84a4e-fc72-4682-87fc-6e7f36204764-host\") pod \"crc-debug-ksgws\" (UID: \"fbc84a4e-fc72-4682-87fc-6e7f36204764\") " pod="openshift-must-gather-k97pd/crc-debug-ksgws" Dec 10 10:31:38 crc kubenswrapper[4715]: I1210 10:31:38.169447 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmhhp\" (UniqueName: \"kubernetes.io/projected/fbc84a4e-fc72-4682-87fc-6e7f36204764-kube-api-access-kmhhp\") pod \"crc-debug-ksgws\" (UID: \"fbc84a4e-fc72-4682-87fc-6e7f36204764\") " pod="openshift-must-gather-k97pd/crc-debug-ksgws" Dec 10 10:31:38 crc kubenswrapper[4715]: I1210 10:31:38.273005 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmhhp\" (UniqueName: \"kubernetes.io/projected/fbc84a4e-fc72-4682-87fc-6e7f36204764-kube-api-access-kmhhp\") pod \"crc-debug-ksgws\" (UID: \"fbc84a4e-fc72-4682-87fc-6e7f36204764\") " pod="openshift-must-gather-k97pd/crc-debug-ksgws" Dec 10 10:31:38 crc kubenswrapper[4715]: I1210 10:31:38.273251 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbc84a4e-fc72-4682-87fc-6e7f36204764-host\") pod \"crc-debug-ksgws\" (UID: \"fbc84a4e-fc72-4682-87fc-6e7f36204764\") " pod="openshift-must-gather-k97pd/crc-debug-ksgws" Dec 10 10:31:38 crc kubenswrapper[4715]: I1210 10:31:38.273459 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbc84a4e-fc72-4682-87fc-6e7f36204764-host\") pod \"crc-debug-ksgws\" (UID: \"fbc84a4e-fc72-4682-87fc-6e7f36204764\") " pod="openshift-must-gather-k97pd/crc-debug-ksgws" Dec 10 10:31:38 crc kubenswrapper[4715]: I1210 10:31:38.311115 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmhhp\" (UniqueName: \"kubernetes.io/projected/fbc84a4e-fc72-4682-87fc-6e7f36204764-kube-api-access-kmhhp\") pod \"crc-debug-ksgws\" (UID: \"fbc84a4e-fc72-4682-87fc-6e7f36204764\") " pod="openshift-must-gather-k97pd/crc-debug-ksgws" Dec 10 10:31:38 crc kubenswrapper[4715]: I1210 10:31:38.411449 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-ksgws" Dec 10 10:31:41 crc kubenswrapper[4715]: I1210 10:31:41.882866 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrt7p" event={"ID":"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c","Type":"ContainerStarted","Data":"124c91881ee6034e556adb16e644c266d9858cd29ea4c3fc11dc5a2d3891d692"} Dec 10 10:31:41 crc kubenswrapper[4715]: I1210 10:31:41.884869 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/crc-debug-ksgws" event={"ID":"fbc84a4e-fc72-4682-87fc-6e7f36204764","Type":"ContainerStarted","Data":"5dc1799f5010d5219b9eb687f5916c58392f5fa9c23d2e1ae4eb777e3e93b80f"} Dec 10 10:31:41 crc kubenswrapper[4715]: I1210 10:31:41.913413 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hrt7p" podStartSLOduration=7.404863034 podStartE2EDuration="13.913375657s" podCreationTimestamp="2025-12-10 10:31:28 +0000 UTC" firstStartedPulling="2025-12-10 10:31:34.815653986 +0000 UTC m=+3457.559200237" lastFinishedPulling="2025-12-10 10:31:41.324166609 +0000 UTC m=+3464.067712860" observedRunningTime="2025-12-10 10:31:41.909867607 +0000 UTC m=+3464.653413858" watchObservedRunningTime="2025-12-10 10:31:41.913375657 +0000 UTC m=+3464.656921908" Dec 10 10:31:48 crc kubenswrapper[4715]: I1210 10:31:48.421457 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:48 crc kubenswrapper[4715]: I1210 10:31:48.422217 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:48 crc kubenswrapper[4715]: I1210 10:31:48.495857 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:49 crc kubenswrapper[4715]: I1210 10:31:49.014598 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:49 crc kubenswrapper[4715]: I1210 10:31:49.069519 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hrt7p"] Dec 10 10:31:50 crc kubenswrapper[4715]: I1210 10:31:50.973145 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hrt7p" podUID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerName="registry-server" containerID="cri-o://124c91881ee6034e556adb16e644c266d9858cd29ea4c3fc11dc5a2d3891d692" gracePeriod=2 Dec 10 10:31:51 crc kubenswrapper[4715]: I1210 10:31:51.988537 4715 generic.go:334] "Generic (PLEG): container finished" podID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerID="124c91881ee6034e556adb16e644c266d9858cd29ea4c3fc11dc5a2d3891d692" exitCode=0 Dec 10 10:31:51 crc kubenswrapper[4715]: I1210 10:31:51.988618 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrt7p" event={"ID":"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c","Type":"ContainerDied","Data":"124c91881ee6034e556adb16e644c266d9858cd29ea4c3fc11dc5a2d3891d692"} Dec 10 10:31:53 crc kubenswrapper[4715]: I1210 10:31:53.189376 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:53 crc kubenswrapper[4715]: I1210 10:31:53.198998 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-catalog-content\") pod \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " Dec 10 10:31:53 crc kubenswrapper[4715]: I1210 10:31:53.199071 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44p5b\" (UniqueName: \"kubernetes.io/projected/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-kube-api-access-44p5b\") pod \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " Dec 10 10:31:53 crc kubenswrapper[4715]: I1210 10:31:53.199168 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-utilities\") pod \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\" (UID: \"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c\") " Dec 10 10:31:53 crc kubenswrapper[4715]: I1210 10:31:53.200153 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-utilities" (OuterVolumeSpecName: "utilities") pod "9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" (UID: "9a4c9658-c2b9-48dd-ad19-2ce27b963a1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:31:53 crc kubenswrapper[4715]: I1210 10:31:53.210880 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-kube-api-access-44p5b" (OuterVolumeSpecName: "kube-api-access-44p5b") pod "9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" (UID: "9a4c9658-c2b9-48dd-ad19-2ce27b963a1c"). InnerVolumeSpecName "kube-api-access-44p5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:31:53 crc kubenswrapper[4715]: I1210 10:31:53.287446 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" (UID: "9a4c9658-c2b9-48dd-ad19-2ce27b963a1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:31:53 crc kubenswrapper[4715]: I1210 10:31:53.301603 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:31:53 crc kubenswrapper[4715]: I1210 10:31:53.301638 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:31:53 crc kubenswrapper[4715]: I1210 10:31:53.301651 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44p5b\" (UniqueName: \"kubernetes.io/projected/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c-kube-api-access-44p5b\") on node \"crc\" DevicePath \"\"" Dec 10 10:31:54 crc kubenswrapper[4715]: I1210 10:31:54.012567 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrt7p" event={"ID":"9a4c9658-c2b9-48dd-ad19-2ce27b963a1c","Type":"ContainerDied","Data":"79f2dc56d4b67294e059d6a953b8b56f8f58f5cf1940d9334b2e88f48851282b"} Dec 10 10:31:54 crc kubenswrapper[4715]: I1210 10:31:54.013094 4715 scope.go:117] "RemoveContainer" containerID="124c91881ee6034e556adb16e644c266d9858cd29ea4c3fc11dc5a2d3891d692" Dec 10 10:31:54 crc kubenswrapper[4715]: I1210 10:31:54.012649 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrt7p" Dec 10 10:31:54 crc kubenswrapper[4715]: I1210 10:31:54.052634 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hrt7p"] Dec 10 10:31:54 crc kubenswrapper[4715]: I1210 10:31:54.055307 4715 scope.go:117] "RemoveContainer" containerID="074ac3ad85de30771334e668f7948bec8e97f5cb4f7f42081052d44fc22ee4da" Dec 10 10:31:54 crc kubenswrapper[4715]: I1210 10:31:54.063849 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hrt7p"] Dec 10 10:31:54 crc kubenswrapper[4715]: I1210 10:31:54.085675 4715 scope.go:117] "RemoveContainer" containerID="61b4020ed5f339956ed930c03c4ca46993dcc0496f5c36c2bd498a2c8116d5ab" Dec 10 10:31:55 crc kubenswrapper[4715]: I1210 10:31:55.026027 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/crc-debug-ksgws" event={"ID":"fbc84a4e-fc72-4682-87fc-6e7f36204764","Type":"ContainerStarted","Data":"6ba138c9fd5e85c1854ebd672b13f91bc7455abb9ad7594bba674ffa9cf81015"} Dec 10 10:31:55 crc kubenswrapper[4715]: I1210 10:31:55.056084 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-k97pd/crc-debug-ksgws" podStartSLOduration=5.001010049 podStartE2EDuration="17.056058805s" podCreationTimestamp="2025-12-10 10:31:38 +0000 UTC" firstStartedPulling="2025-12-10 10:31:40.885430341 +0000 UTC m=+3463.628976592" lastFinishedPulling="2025-12-10 10:31:52.940479097 +0000 UTC m=+3475.684025348" observedRunningTime="2025-12-10 10:31:55.046881133 +0000 UTC m=+3477.790427384" watchObservedRunningTime="2025-12-10 10:31:55.056058805 +0000 UTC m=+3477.799605066" Dec 10 10:31:55 crc kubenswrapper[4715]: I1210 10:31:55.617257 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" path="/var/lib/kubelet/pods/9a4c9658-c2b9-48dd-ad19-2ce27b963a1c/volumes" Dec 10 10:32:47 crc kubenswrapper[4715]: I1210 10:32:47.567494 4715 generic.go:334] "Generic (PLEG): container finished" podID="fbc84a4e-fc72-4682-87fc-6e7f36204764" containerID="6ba138c9fd5e85c1854ebd672b13f91bc7455abb9ad7594bba674ffa9cf81015" exitCode=0 Dec 10 10:32:47 crc kubenswrapper[4715]: I1210 10:32:47.567606 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/crc-debug-ksgws" event={"ID":"fbc84a4e-fc72-4682-87fc-6e7f36204764","Type":"ContainerDied","Data":"6ba138c9fd5e85c1854ebd672b13f91bc7455abb9ad7594bba674ffa9cf81015"} Dec 10 10:32:47 crc kubenswrapper[4715]: I1210 10:32:47.713736 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:32:47 crc kubenswrapper[4715]: I1210 10:32:47.713811 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:32:48 crc kubenswrapper[4715]: I1210 10:32:48.684831 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-ksgws" Dec 10 10:32:48 crc kubenswrapper[4715]: I1210 10:32:48.735320 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-k97pd/crc-debug-ksgws"] Dec 10 10:32:48 crc kubenswrapper[4715]: I1210 10:32:48.744341 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-k97pd/crc-debug-ksgws"] Dec 10 10:32:48 crc kubenswrapper[4715]: I1210 10:32:48.786099 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbc84a4e-fc72-4682-87fc-6e7f36204764-host\") pod \"fbc84a4e-fc72-4682-87fc-6e7f36204764\" (UID: \"fbc84a4e-fc72-4682-87fc-6e7f36204764\") " Dec 10 10:32:48 crc kubenswrapper[4715]: I1210 10:32:48.786431 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmhhp\" (UniqueName: \"kubernetes.io/projected/fbc84a4e-fc72-4682-87fc-6e7f36204764-kube-api-access-kmhhp\") pod \"fbc84a4e-fc72-4682-87fc-6e7f36204764\" (UID: \"fbc84a4e-fc72-4682-87fc-6e7f36204764\") " Dec 10 10:32:48 crc kubenswrapper[4715]: I1210 10:32:48.786537 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fbc84a4e-fc72-4682-87fc-6e7f36204764-host" (OuterVolumeSpecName: "host") pod "fbc84a4e-fc72-4682-87fc-6e7f36204764" (UID: "fbc84a4e-fc72-4682-87fc-6e7f36204764"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 10:32:48 crc kubenswrapper[4715]: I1210 10:32:48.786847 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbc84a4e-fc72-4682-87fc-6e7f36204764-host\") on node \"crc\" DevicePath \"\"" Dec 10 10:32:48 crc kubenswrapper[4715]: I1210 10:32:48.792803 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbc84a4e-fc72-4682-87fc-6e7f36204764-kube-api-access-kmhhp" (OuterVolumeSpecName: "kube-api-access-kmhhp") pod "fbc84a4e-fc72-4682-87fc-6e7f36204764" (UID: "fbc84a4e-fc72-4682-87fc-6e7f36204764"). InnerVolumeSpecName "kube-api-access-kmhhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:32:48 crc kubenswrapper[4715]: I1210 10:32:48.888274 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmhhp\" (UniqueName: \"kubernetes.io/projected/fbc84a4e-fc72-4682-87fc-6e7f36204764-kube-api-access-kmhhp\") on node \"crc\" DevicePath \"\"" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.590798 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dc1799f5010d5219b9eb687f5916c58392f5fa9c23d2e1ae4eb777e3e93b80f" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.590880 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-ksgws" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.627843 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbc84a4e-fc72-4682-87fc-6e7f36204764" path="/var/lib/kubelet/pods/fbc84a4e-fc72-4682-87fc-6e7f36204764/volumes" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.902024 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-k97pd/crc-debug-ft8mh"] Dec 10 10:32:49 crc kubenswrapper[4715]: E1210 10:32:49.902439 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerName="registry-server" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.902452 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerName="registry-server" Dec 10 10:32:49 crc kubenswrapper[4715]: E1210 10:32:49.902474 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbc84a4e-fc72-4682-87fc-6e7f36204764" containerName="container-00" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.902482 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbc84a4e-fc72-4682-87fc-6e7f36204764" containerName="container-00" Dec 10 10:32:49 crc kubenswrapper[4715]: E1210 10:32:49.902498 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerName="extract-utilities" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.902508 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerName="extract-utilities" Dec 10 10:32:49 crc kubenswrapper[4715]: E1210 10:32:49.902539 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerName="extract-content" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.902546 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerName="extract-content" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.902770 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4c9658-c2b9-48dd-ad19-2ce27b963a1c" containerName="registry-server" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.902808 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbc84a4e-fc72-4682-87fc-6e7f36204764" containerName="container-00" Dec 10 10:32:49 crc kubenswrapper[4715]: I1210 10:32:49.903494 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-ft8mh" Dec 10 10:32:50 crc kubenswrapper[4715]: I1210 10:32:50.009868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tspd6\" (UniqueName: \"kubernetes.io/projected/9252c211-2dae-4b92-b494-8529abdb46aa-kube-api-access-tspd6\") pod \"crc-debug-ft8mh\" (UID: \"9252c211-2dae-4b92-b494-8529abdb46aa\") " pod="openshift-must-gather-k97pd/crc-debug-ft8mh" Dec 10 10:32:50 crc kubenswrapper[4715]: I1210 10:32:50.009960 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9252c211-2dae-4b92-b494-8529abdb46aa-host\") pod \"crc-debug-ft8mh\" (UID: \"9252c211-2dae-4b92-b494-8529abdb46aa\") " pod="openshift-must-gather-k97pd/crc-debug-ft8mh" Dec 10 10:32:50 crc kubenswrapper[4715]: I1210 10:32:50.112028 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tspd6\" (UniqueName: \"kubernetes.io/projected/9252c211-2dae-4b92-b494-8529abdb46aa-kube-api-access-tspd6\") pod \"crc-debug-ft8mh\" (UID: \"9252c211-2dae-4b92-b494-8529abdb46aa\") " pod="openshift-must-gather-k97pd/crc-debug-ft8mh" Dec 10 10:32:50 crc kubenswrapper[4715]: I1210 10:32:50.112145 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9252c211-2dae-4b92-b494-8529abdb46aa-host\") pod \"crc-debug-ft8mh\" (UID: \"9252c211-2dae-4b92-b494-8529abdb46aa\") " pod="openshift-must-gather-k97pd/crc-debug-ft8mh" Dec 10 10:32:50 crc kubenswrapper[4715]: I1210 10:32:50.112312 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9252c211-2dae-4b92-b494-8529abdb46aa-host\") pod \"crc-debug-ft8mh\" (UID: \"9252c211-2dae-4b92-b494-8529abdb46aa\") " pod="openshift-must-gather-k97pd/crc-debug-ft8mh" Dec 10 10:32:50 crc kubenswrapper[4715]: I1210 10:32:50.136939 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tspd6\" (UniqueName: \"kubernetes.io/projected/9252c211-2dae-4b92-b494-8529abdb46aa-kube-api-access-tspd6\") pod \"crc-debug-ft8mh\" (UID: \"9252c211-2dae-4b92-b494-8529abdb46aa\") " pod="openshift-must-gather-k97pd/crc-debug-ft8mh" Dec 10 10:32:50 crc kubenswrapper[4715]: I1210 10:32:50.226701 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-ft8mh" Dec 10 10:32:50 crc kubenswrapper[4715]: I1210 10:32:50.602383 4715 generic.go:334] "Generic (PLEG): container finished" podID="9252c211-2dae-4b92-b494-8529abdb46aa" containerID="88ea0e2a1ce564f9a4192e86f9c24710f683e04ca66eb4030c55558f40351551" exitCode=0 Dec 10 10:32:50 crc kubenswrapper[4715]: I1210 10:32:50.602471 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/crc-debug-ft8mh" event={"ID":"9252c211-2dae-4b92-b494-8529abdb46aa","Type":"ContainerDied","Data":"88ea0e2a1ce564f9a4192e86f9c24710f683e04ca66eb4030c55558f40351551"} Dec 10 10:32:50 crc kubenswrapper[4715]: I1210 10:32:50.602719 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/crc-debug-ft8mh" event={"ID":"9252c211-2dae-4b92-b494-8529abdb46aa","Type":"ContainerStarted","Data":"8bbba765a1edb48e59f56b63cb2f7701e541a150e2c3edbe869405c6e2433d1b"} Dec 10 10:32:51 crc kubenswrapper[4715]: I1210 10:32:51.126561 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-k97pd/crc-debug-ft8mh"] Dec 10 10:32:51 crc kubenswrapper[4715]: I1210 10:32:51.135046 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-k97pd/crc-debug-ft8mh"] Dec 10 10:32:51 crc kubenswrapper[4715]: I1210 10:32:51.749524 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-ft8mh" Dec 10 10:32:51 crc kubenswrapper[4715]: I1210 10:32:51.848323 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9252c211-2dae-4b92-b494-8529abdb46aa-host\") pod \"9252c211-2dae-4b92-b494-8529abdb46aa\" (UID: \"9252c211-2dae-4b92-b494-8529abdb46aa\") " Dec 10 10:32:51 crc kubenswrapper[4715]: I1210 10:32:51.848538 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9252c211-2dae-4b92-b494-8529abdb46aa-host" (OuterVolumeSpecName: "host") pod "9252c211-2dae-4b92-b494-8529abdb46aa" (UID: "9252c211-2dae-4b92-b494-8529abdb46aa"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 10:32:51 crc kubenswrapper[4715]: I1210 10:32:51.848638 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tspd6\" (UniqueName: \"kubernetes.io/projected/9252c211-2dae-4b92-b494-8529abdb46aa-kube-api-access-tspd6\") pod \"9252c211-2dae-4b92-b494-8529abdb46aa\" (UID: \"9252c211-2dae-4b92-b494-8529abdb46aa\") " Dec 10 10:32:51 crc kubenswrapper[4715]: I1210 10:32:51.849443 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9252c211-2dae-4b92-b494-8529abdb46aa-host\") on node \"crc\" DevicePath \"\"" Dec 10 10:32:51 crc kubenswrapper[4715]: I1210 10:32:51.858194 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9252c211-2dae-4b92-b494-8529abdb46aa-kube-api-access-tspd6" (OuterVolumeSpecName: "kube-api-access-tspd6") pod "9252c211-2dae-4b92-b494-8529abdb46aa" (UID: "9252c211-2dae-4b92-b494-8529abdb46aa"). InnerVolumeSpecName "kube-api-access-tspd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:32:51 crc kubenswrapper[4715]: I1210 10:32:51.951550 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tspd6\" (UniqueName: \"kubernetes.io/projected/9252c211-2dae-4b92-b494-8529abdb46aa-kube-api-access-tspd6\") on node \"crc\" DevicePath \"\"" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.353602 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-k97pd/crc-debug-7vb9w"] Dec 10 10:32:52 crc kubenswrapper[4715]: E1210 10:32:52.354431 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9252c211-2dae-4b92-b494-8529abdb46aa" containerName="container-00" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.354444 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9252c211-2dae-4b92-b494-8529abdb46aa" containerName="container-00" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.354655 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9252c211-2dae-4b92-b494-8529abdb46aa" containerName="container-00" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.355429 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-7vb9w" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.462708 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-host\") pod \"crc-debug-7vb9w\" (UID: \"ac074277-1f6f-4d4a-9b2a-925d7c5646f6\") " pod="openshift-must-gather-k97pd/crc-debug-7vb9w" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.462896 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5rtp\" (UniqueName: \"kubernetes.io/projected/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-kube-api-access-q5rtp\") pod \"crc-debug-7vb9w\" (UID: \"ac074277-1f6f-4d4a-9b2a-925d7c5646f6\") " pod="openshift-must-gather-k97pd/crc-debug-7vb9w" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.565038 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5rtp\" (UniqueName: \"kubernetes.io/projected/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-kube-api-access-q5rtp\") pod \"crc-debug-7vb9w\" (UID: \"ac074277-1f6f-4d4a-9b2a-925d7c5646f6\") " pod="openshift-must-gather-k97pd/crc-debug-7vb9w" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.565207 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-host\") pod \"crc-debug-7vb9w\" (UID: \"ac074277-1f6f-4d4a-9b2a-925d7c5646f6\") " pod="openshift-must-gather-k97pd/crc-debug-7vb9w" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.565322 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-host\") pod \"crc-debug-7vb9w\" (UID: \"ac074277-1f6f-4d4a-9b2a-925d7c5646f6\") " pod="openshift-must-gather-k97pd/crc-debug-7vb9w" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.589712 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5rtp\" (UniqueName: \"kubernetes.io/projected/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-kube-api-access-q5rtp\") pod \"crc-debug-7vb9w\" (UID: \"ac074277-1f6f-4d4a-9b2a-925d7c5646f6\") " pod="openshift-must-gather-k97pd/crc-debug-7vb9w" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.625562 4715 scope.go:117] "RemoveContainer" containerID="88ea0e2a1ce564f9a4192e86f9c24710f683e04ca66eb4030c55558f40351551" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.625679 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-ft8mh" Dec 10 10:32:52 crc kubenswrapper[4715]: I1210 10:32:52.674665 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-7vb9w" Dec 10 10:32:52 crc kubenswrapper[4715]: W1210 10:32:52.722094 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac074277_1f6f_4d4a_9b2a_925d7c5646f6.slice/crio-30cd225d5dd36efa05df2a481baa7dd3dd422d884a10176232e520fd5fbf1d7c WatchSource:0}: Error finding container 30cd225d5dd36efa05df2a481baa7dd3dd422d884a10176232e520fd5fbf1d7c: Status 404 returned error can't find the container with id 30cd225d5dd36efa05df2a481baa7dd3dd422d884a10176232e520fd5fbf1d7c Dec 10 10:32:53 crc kubenswrapper[4715]: I1210 10:32:53.616860 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9252c211-2dae-4b92-b494-8529abdb46aa" path="/var/lib/kubelet/pods/9252c211-2dae-4b92-b494-8529abdb46aa/volumes" Dec 10 10:32:53 crc kubenswrapper[4715]: I1210 10:32:53.636394 4715 generic.go:334] "Generic (PLEG): container finished" podID="ac074277-1f6f-4d4a-9b2a-925d7c5646f6" containerID="a4800f20a509e9c97ed2e98a97e597740a0de5b4881ef8151015847909d2d604" exitCode=0 Dec 10 10:32:53 crc kubenswrapper[4715]: I1210 10:32:53.636471 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/crc-debug-7vb9w" event={"ID":"ac074277-1f6f-4d4a-9b2a-925d7c5646f6","Type":"ContainerDied","Data":"a4800f20a509e9c97ed2e98a97e597740a0de5b4881ef8151015847909d2d604"} Dec 10 10:32:53 crc kubenswrapper[4715]: I1210 10:32:53.636832 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/crc-debug-7vb9w" event={"ID":"ac074277-1f6f-4d4a-9b2a-925d7c5646f6","Type":"ContainerStarted","Data":"30cd225d5dd36efa05df2a481baa7dd3dd422d884a10176232e520fd5fbf1d7c"} Dec 10 10:32:53 crc kubenswrapper[4715]: I1210 10:32:53.678299 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-k97pd/crc-debug-7vb9w"] Dec 10 10:32:53 crc kubenswrapper[4715]: I1210 10:32:53.689270 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-k97pd/crc-debug-7vb9w"] Dec 10 10:32:54 crc kubenswrapper[4715]: I1210 10:32:54.751576 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-7vb9w" Dec 10 10:32:54 crc kubenswrapper[4715]: I1210 10:32:54.845512 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-host\") pod \"ac074277-1f6f-4d4a-9b2a-925d7c5646f6\" (UID: \"ac074277-1f6f-4d4a-9b2a-925d7c5646f6\") " Dec 10 10:32:54 crc kubenswrapper[4715]: I1210 10:32:54.845659 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-host" (OuterVolumeSpecName: "host") pod "ac074277-1f6f-4d4a-9b2a-925d7c5646f6" (UID: "ac074277-1f6f-4d4a-9b2a-925d7c5646f6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 10:32:54 crc kubenswrapper[4715]: I1210 10:32:54.845836 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5rtp\" (UniqueName: \"kubernetes.io/projected/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-kube-api-access-q5rtp\") pod \"ac074277-1f6f-4d4a-9b2a-925d7c5646f6\" (UID: \"ac074277-1f6f-4d4a-9b2a-925d7c5646f6\") " Dec 10 10:32:54 crc kubenswrapper[4715]: I1210 10:32:54.846381 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-host\") on node \"crc\" DevicePath \"\"" Dec 10 10:32:54 crc kubenswrapper[4715]: I1210 10:32:54.852904 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-kube-api-access-q5rtp" (OuterVolumeSpecName: "kube-api-access-q5rtp") pod "ac074277-1f6f-4d4a-9b2a-925d7c5646f6" (UID: "ac074277-1f6f-4d4a-9b2a-925d7c5646f6"). InnerVolumeSpecName "kube-api-access-q5rtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:32:54 crc kubenswrapper[4715]: I1210 10:32:54.948873 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5rtp\" (UniqueName: \"kubernetes.io/projected/ac074277-1f6f-4d4a-9b2a-925d7c5646f6-kube-api-access-q5rtp\") on node \"crc\" DevicePath \"\"" Dec 10 10:32:55 crc kubenswrapper[4715]: I1210 10:32:55.615184 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac074277-1f6f-4d4a-9b2a-925d7c5646f6" path="/var/lib/kubelet/pods/ac074277-1f6f-4d4a-9b2a-925d7c5646f6/volumes" Dec 10 10:32:55 crc kubenswrapper[4715]: I1210 10:32:55.659486 4715 scope.go:117] "RemoveContainer" containerID="a4800f20a509e9c97ed2e98a97e597740a0de5b4881ef8151015847909d2d604" Dec 10 10:32:55 crc kubenswrapper[4715]: I1210 10:32:55.659585 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/crc-debug-7vb9w" Dec 10 10:33:11 crc kubenswrapper[4715]: I1210 10:33:11.020062 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7856564b58-9kr94_07a44832-3242-42dd-9366-5628181a6c6e/barbican-api/0.log" Dec 10 10:33:11 crc kubenswrapper[4715]: I1210 10:33:11.240662 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7856564b58-9kr94_07a44832-3242-42dd-9366-5628181a6c6e/barbican-api-log/0.log" Dec 10 10:33:11 crc kubenswrapper[4715]: I1210 10:33:11.285299 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6d588649f8-m9jx2_73357509-8bf3-408e-9c35-ee6267593be1/barbican-keystone-listener/0.log" Dec 10 10:33:11 crc kubenswrapper[4715]: I1210 10:33:11.314045 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6d588649f8-m9jx2_73357509-8bf3-408e-9c35-ee6267593be1/barbican-keystone-listener-log/0.log" Dec 10 10:33:11 crc kubenswrapper[4715]: I1210 10:33:11.464678 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c964fbd47-67rln_b46f5949-5bfa-4cfd-b70d-4d225ea12d6b/barbican-worker/0.log" Dec 10 10:33:11 crc kubenswrapper[4715]: I1210 10:33:11.503664 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c964fbd47-67rln_b46f5949-5bfa-4cfd-b70d-4d225ea12d6b/barbican-worker-log/0.log" Dec 10 10:33:11 crc kubenswrapper[4715]: I1210 10:33:11.939532 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3a26e430-a5f3-40cf-97fc-be1e01adb4e9/proxy-httpd/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.003782 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3a26e430-a5f3-40cf-97fc-be1e01adb4e9/ceilometer-notification-agent/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.008531 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw_7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.016670 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3a26e430-a5f3-40cf-97fc-be1e01adb4e9/ceilometer-central-agent/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.127000 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3a26e430-a5f3-40cf-97fc-be1e01adb4e9/sg-core/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.253313 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e911669d-5a8e-49d6-a6e2-b66dfae8f7e5/cinder-api-log/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.253759 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e911669d-5a8e-49d6-a6e2-b66dfae8f7e5/cinder-api/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.472828 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_807957b7-0d34-49ab-8b79-6abcd56ad840/cinder-scheduler/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.506707 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_807957b7-0d34-49ab-8b79-6abcd56ad840/probe/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.592119 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl_42754cce-1e7d-416d-a1fa-118ae307880f/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.720629 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-rxk96_c952af8b-0155-480f-a60f-399b9ec919c8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.809207 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-g755x_540bd567-e127-4db2-b6c5-0d447ce7b34c/init/0.log" Dec 10 10:33:12 crc kubenswrapper[4715]: I1210 10:33:12.999586 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-g755x_540bd567-e127-4db2-b6c5-0d447ce7b34c/init/0.log" Dec 10 10:33:13 crc kubenswrapper[4715]: I1210 10:33:13.066501 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-g755x_540bd567-e127-4db2-b6c5-0d447ce7b34c/dnsmasq-dns/0.log" Dec 10 10:33:13 crc kubenswrapper[4715]: I1210 10:33:13.086057 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x_6d7bc0b2-e15d-4627-b901-bb6399f7f636/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:13 crc kubenswrapper[4715]: I1210 10:33:13.259472 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac/glance-httpd/0.log" Dec 10 10:33:13 crc kubenswrapper[4715]: I1210 10:33:13.320703 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac/glance-log/0.log" Dec 10 10:33:13 crc kubenswrapper[4715]: I1210 10:33:13.435132 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ba040a24-bae6-4627-a934-39d616b57c76/glance-log/0.log" Dec 10 10:33:13 crc kubenswrapper[4715]: I1210 10:33:13.495480 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ba040a24-bae6-4627-a934-39d616b57c76/glance-httpd/0.log" Dec 10 10:33:13 crc kubenswrapper[4715]: I1210 10:33:13.645621 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-857b4dfcd4-wnlbz_4214260f-423f-47fb-b062-c752f3519175/horizon/0.log" Dec 10 10:33:13 crc kubenswrapper[4715]: I1210 10:33:13.895315 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk_0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:14 crc kubenswrapper[4715]: I1210 10:33:14.039749 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-857b4dfcd4-wnlbz_4214260f-423f-47fb-b062-c752f3519175/horizon-log/0.log" Dec 10 10:33:14 crc kubenswrapper[4715]: I1210 10:33:14.072134 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-h82b4_9692f891-f3d4-412b-92bd-d154842e0d79/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:14 crc kubenswrapper[4715]: I1210 10:33:14.139262 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29422681-fb8s5_2992fa3e-b3d4-42da-845e-a019e8e23cf1/keystone-cron/0.log" Dec 10 10:33:14 crc kubenswrapper[4715]: I1210 10:33:14.411888 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_cf760357-b816-4b56-b56d-5a3637a7b2b0/kube-state-metrics/0.log" Dec 10 10:33:14 crc kubenswrapper[4715]: I1210 10:33:14.500288 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-d9c5c96bf-n94nb_946c4b3b-eb91-4d39-a7ef-88d25e23599e/keystone-api/0.log" Dec 10 10:33:14 crc kubenswrapper[4715]: I1210 10:33:14.686756 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-959dc_cd86f221-0c52-4565-87ed-9d4c473aaaf4/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:15 crc kubenswrapper[4715]: I1210 10:33:15.058288 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-85df69bf85-6cwpq_0cf58c4f-3594-4b20-a3fb-313c0f25a844/neutron-httpd/0.log" Dec 10 10:33:15 crc kubenswrapper[4715]: I1210 10:33:15.081740 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-85df69bf85-6cwpq_0cf58c4f-3594-4b20-a3fb-313c0f25a844/neutron-api/0.log" Dec 10 10:33:15 crc kubenswrapper[4715]: I1210 10:33:15.322048 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6_06251a0c-f042-4137-864b-c7003449b040/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:15 crc kubenswrapper[4715]: I1210 10:33:15.801059 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2648041d-c409-4012-b90b-3d780bacb44d/nova-api-log/0.log" Dec 10 10:33:15 crc kubenswrapper[4715]: I1210 10:33:15.823273 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da/nova-cell0-conductor-conductor/0.log" Dec 10 10:33:15 crc kubenswrapper[4715]: I1210 10:33:15.970425 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2648041d-c409-4012-b90b-3d780bacb44d/nova-api-api/0.log" Dec 10 10:33:16 crc kubenswrapper[4715]: I1210 10:33:16.094425 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7748551a-94b0-47c5-8360-584b386b0bf2/nova-cell1-conductor-conductor/0.log" Dec 10 10:33:16 crc kubenswrapper[4715]: I1210 10:33:16.259139 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_87397cd7-eabb-4b0b-809a-c9f2ab53e8be/nova-cell1-novncproxy-novncproxy/0.log" Dec 10 10:33:16 crc kubenswrapper[4715]: I1210 10:33:16.402523 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-sbpnp_31ba5b33-f7ca-4389-879d-8a6a38cc0795/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:16 crc kubenswrapper[4715]: I1210 10:33:16.581464 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8027e8a4-9be7-4764-8c1d-f25a3eb3587b/nova-metadata-log/0.log" Dec 10 10:33:16 crc kubenswrapper[4715]: I1210 10:33:16.780001 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_6feb8a72-37ea-419b-b69e-0a2ef81d79ca/nova-scheduler-scheduler/0.log" Dec 10 10:33:16 crc kubenswrapper[4715]: I1210 10:33:16.878412 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_05dd65e3-40ea-4d12-856f-3fa4f99712b2/mysql-bootstrap/0.log" Dec 10 10:33:17 crc kubenswrapper[4715]: I1210 10:33:17.087423 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_05dd65e3-40ea-4d12-856f-3fa4f99712b2/galera/0.log" Dec 10 10:33:17 crc kubenswrapper[4715]: I1210 10:33:17.182943 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_05dd65e3-40ea-4d12-856f-3fa4f99712b2/mysql-bootstrap/0.log" Dec 10 10:33:17 crc kubenswrapper[4715]: I1210 10:33:17.323384 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8bcd31bd-4e40-41b6-92ef-6e0e0b170afa/mysql-bootstrap/0.log" Dec 10 10:33:17 crc kubenswrapper[4715]: I1210 10:33:17.607758 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8bcd31bd-4e40-41b6-92ef-6e0e0b170afa/mysql-bootstrap/0.log" Dec 10 10:33:17 crc kubenswrapper[4715]: I1210 10:33:17.661483 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8bcd31bd-4e40-41b6-92ef-6e0e0b170afa/galera/0.log" Dec 10 10:33:17 crc kubenswrapper[4715]: I1210 10:33:17.714462 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:33:17 crc kubenswrapper[4715]: I1210 10:33:17.714531 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:33:17 crc kubenswrapper[4715]: I1210 10:33:17.803987 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2843e1e0-1b99-45af-a422-e533ee493c78/openstackclient/0.log" Dec 10 10:33:17 crc kubenswrapper[4715]: I1210 10:33:17.854094 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8027e8a4-9be7-4764-8c1d-f25a3eb3587b/nova-metadata-metadata/0.log" Dec 10 10:33:17 crc kubenswrapper[4715]: I1210 10:33:17.941102 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-lz7jm_6fc17771-af47-4ba4-be8a-6b3143a999f4/ovn-controller/0.log" Dec 10 10:33:18 crc kubenswrapper[4715]: I1210 10:33:18.130813 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-p5m86_e26b3d5b-1e63-4d22-80a5-55ffb58f664b/openstack-network-exporter/0.log" Dec 10 10:33:18 crc kubenswrapper[4715]: I1210 10:33:18.421928 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rv2hq_d6ae191b-a288-4185-b98b-15654b1e5dbf/ovsdb-server-init/0.log" Dec 10 10:33:18 crc kubenswrapper[4715]: I1210 10:33:18.633987 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rv2hq_d6ae191b-a288-4185-b98b-15654b1e5dbf/ovsdb-server-init/0.log" Dec 10 10:33:18 crc kubenswrapper[4715]: I1210 10:33:18.664850 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rv2hq_d6ae191b-a288-4185-b98b-15654b1e5dbf/ovsdb-server/0.log" Dec 10 10:33:18 crc kubenswrapper[4715]: I1210 10:33:18.677691 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rv2hq_d6ae191b-a288-4185-b98b-15654b1e5dbf/ovs-vswitchd/0.log" Dec 10 10:33:18 crc kubenswrapper[4715]: I1210 10:33:18.892815 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-plxxq_b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:18 crc kubenswrapper[4715]: I1210 10:33:18.916257 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b0f6351-5ab5-4e0c-ad34-95061bc14a35/openstack-network-exporter/0.log" Dec 10 10:33:19 crc kubenswrapper[4715]: I1210 10:33:19.110194 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b0f6351-5ab5-4e0c-ad34-95061bc14a35/ovn-northd/0.log" Dec 10 10:33:19 crc kubenswrapper[4715]: I1210 10:33:19.180422 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b4bfd970-df85-4d59-a211-ba4adbb501ee/openstack-network-exporter/0.log" Dec 10 10:33:19 crc kubenswrapper[4715]: I1210 10:33:19.325098 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b4bfd970-df85-4d59-a211-ba4adbb501ee/ovsdbserver-nb/0.log" Dec 10 10:33:19 crc kubenswrapper[4715]: I1210 10:33:19.463817 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_219b8c44-8f39-40f1-9000-25404c35d495/ovsdbserver-sb/0.log" Dec 10 10:33:19 crc kubenswrapper[4715]: I1210 10:33:19.488076 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_219b8c44-8f39-40f1-9000-25404c35d495/openstack-network-exporter/0.log" Dec 10 10:33:19 crc kubenswrapper[4715]: I1210 10:33:19.765017 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5777d57b56-6lbcp_21593c1a-bc1a-4460-9ef9-3eef12261355/placement-api/0.log" Dec 10 10:33:19 crc kubenswrapper[4715]: I1210 10:33:19.844434 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5777d57b56-6lbcp_21593c1a-bc1a-4460-9ef9-3eef12261355/placement-log/0.log" Dec 10 10:33:19 crc kubenswrapper[4715]: I1210 10:33:19.959328 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf6962bf-0a77-4285-b58f-6213688e74b7/setup-container/0.log" Dec 10 10:33:20 crc kubenswrapper[4715]: I1210 10:33:20.158134 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf6962bf-0a77-4285-b58f-6213688e74b7/setup-container/0.log" Dec 10 10:33:20 crc kubenswrapper[4715]: I1210 10:33:20.251277 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf6962bf-0a77-4285-b58f-6213688e74b7/rabbitmq/0.log" Dec 10 10:33:20 crc kubenswrapper[4715]: I1210 10:33:20.274870 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4e8933c-d768-48cb-bac7-245abd8761ad/setup-container/0.log" Dec 10 10:33:20 crc kubenswrapper[4715]: I1210 10:33:20.515894 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4e8933c-d768-48cb-bac7-245abd8761ad/rabbitmq/0.log" Dec 10 10:33:20 crc kubenswrapper[4715]: I1210 10:33:20.531298 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4e8933c-d768-48cb-bac7-245abd8761ad/setup-container/0.log" Dec 10 10:33:20 crc kubenswrapper[4715]: I1210 10:33:20.540637 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-sr547_6d1a7152-aa86-4400-b3bb-55ca096ef3f7/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:20 crc kubenswrapper[4715]: I1210 10:33:20.796254 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-5rcr7_bd4fa444-6ad3-4129-b62d-c98ce1b69058/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:20 crc kubenswrapper[4715]: I1210 10:33:20.811375 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl_17c08b14-a1de-4e79-b6d6-a8543bb0a4fd/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:21 crc kubenswrapper[4715]: I1210 10:33:21.034120 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pz4s4_06a8246c-8107-4c74-b827-1c965bd840ec/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:21 crc kubenswrapper[4715]: I1210 10:33:21.056334 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8h9tz_757f9102-c847-4fa9-b9a6-87495185cffe/ssh-known-hosts-edpm-deployment/0.log" Dec 10 10:33:21 crc kubenswrapper[4715]: I1210 10:33:21.632299 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6bcc74b4c5-ws76b_34a9b314-68fd-43d1-89c8-b55bf717f6df/proxy-server/0.log" Dec 10 10:33:21 crc kubenswrapper[4715]: I1210 10:33:21.742363 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-47zsh_3a351985-c466-4a2b-b74e-f67d85fb5715/swift-ring-rebalance/0.log" Dec 10 10:33:21 crc kubenswrapper[4715]: I1210 10:33:21.753823 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6bcc74b4c5-ws76b_34a9b314-68fd-43d1-89c8-b55bf717f6df/proxy-httpd/0.log" Dec 10 10:33:21 crc kubenswrapper[4715]: I1210 10:33:21.827285 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/account-auditor/0.log" Dec 10 10:33:21 crc kubenswrapper[4715]: I1210 10:33:21.960712 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/account-reaper/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.014034 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/account-replicator/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.094613 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/account-server/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.126134 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/container-auditor/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.243519 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/container-replicator/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.311384 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/container-server/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.331921 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/object-auditor/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.343128 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/container-updater/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.450094 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/object-expirer/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.562032 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/object-replicator/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.572705 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/object-server/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.573763 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/object-updater/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.668797 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/rsync/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.804579 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/swift-recon-cron/0.log" Dec 10 10:33:22 crc kubenswrapper[4715]: I1210 10:33:22.884347 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk_54173a4c-eb6d-4482-9bce-3f24ab7a8b4f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:23 crc kubenswrapper[4715]: I1210 10:33:23.080049 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_40f62965-1d20-4eba-803a-b513c07e489f/test-operator-logs-container/0.log" Dec 10 10:33:23 crc kubenswrapper[4715]: I1210 10:33:23.132829 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8/tempest-tests-tempest-tests-runner/0.log" Dec 10 10:33:23 crc kubenswrapper[4715]: I1210 10:33:23.301856 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m_dc964814-44b2-4f7c-b4f2-7a9962758553/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:33:31 crc kubenswrapper[4715]: I1210 10:33:31.844738 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_92be40ca-d175-4ec0-819b-1579787701d7/memcached/0.log" Dec 10 10:33:47 crc kubenswrapper[4715]: I1210 10:33:47.714571 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:33:47 crc kubenswrapper[4715]: I1210 10:33:47.715108 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:33:47 crc kubenswrapper[4715]: I1210 10:33:47.715162 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 10:33:47 crc kubenswrapper[4715]: I1210 10:33:47.715988 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"62c20bae2d69f89d01d1894e46f3b38c45cea4c280173ffb7123e37e92db2bf8"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 10:33:47 crc kubenswrapper[4715]: I1210 10:33:47.716044 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://62c20bae2d69f89d01d1894e46f3b38c45cea4c280173ffb7123e37e92db2bf8" gracePeriod=600 Dec 10 10:33:48 crc kubenswrapper[4715]: I1210 10:33:48.297927 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="62c20bae2d69f89d01d1894e46f3b38c45cea4c280173ffb7123e37e92db2bf8" exitCode=0 Dec 10 10:33:48 crc kubenswrapper[4715]: I1210 10:33:48.297989 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"62c20bae2d69f89d01d1894e46f3b38c45cea4c280173ffb7123e37e92db2bf8"} Dec 10 10:33:48 crc kubenswrapper[4715]: I1210 10:33:48.298089 4715 scope.go:117] "RemoveContainer" containerID="6e988da5234f2d2ab05239b7493c957cafd8a1f9635414c6ab59d9d23c706ea2" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.310616 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5"} Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.359194 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t4fdt"] Dec 10 10:33:49 crc kubenswrapper[4715]: E1210 10:33:49.359687 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac074277-1f6f-4d4a-9b2a-925d7c5646f6" containerName="container-00" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.359702 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac074277-1f6f-4d4a-9b2a-925d7c5646f6" containerName="container-00" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.359900 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac074277-1f6f-4d4a-9b2a-925d7c5646f6" containerName="container-00" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.361441 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.375139 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4fdt"] Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.502673 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-catalog-content\") pod \"redhat-marketplace-t4fdt\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.502821 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-utilities\") pod \"redhat-marketplace-t4fdt\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.502871 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn6fl\" (UniqueName: \"kubernetes.io/projected/c945121a-84b7-4a85-ae5c-ad652ceba703-kube-api-access-hn6fl\") pod \"redhat-marketplace-t4fdt\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.604178 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-utilities\") pod \"redhat-marketplace-t4fdt\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.604442 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn6fl\" (UniqueName: \"kubernetes.io/projected/c945121a-84b7-4a85-ae5c-ad652ceba703-kube-api-access-hn6fl\") pod \"redhat-marketplace-t4fdt\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.604571 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-catalog-content\") pod \"redhat-marketplace-t4fdt\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.604732 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-utilities\") pod \"redhat-marketplace-t4fdt\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.604985 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-catalog-content\") pod \"redhat-marketplace-t4fdt\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.624888 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn6fl\" (UniqueName: \"kubernetes.io/projected/c945121a-84b7-4a85-ae5c-ad652ceba703-kube-api-access-hn6fl\") pod \"redhat-marketplace-t4fdt\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.685267 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:49 crc kubenswrapper[4715]: I1210 10:33:49.818477 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/util/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.068598 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/util/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.111977 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/pull/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.171258 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/pull/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.271086 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4fdt"] Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.323373 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4fdt" event={"ID":"c945121a-84b7-4a85-ae5c-ad652ceba703","Type":"ContainerStarted","Data":"39b451f508a376232b8f084a26671360051275255b0bcd41f2ad1e880506ec1e"} Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.376799 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/util/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.434234 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/pull/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.441298 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/extract/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.567054 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jwwm5_f1e4cba4-0694-4477-b031-5caac1fbcb84/kube-rbac-proxy/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.663232 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jwwm5_f1e4cba4-0694-4477-b031-5caac1fbcb84/manager/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.677781 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-htklg_af548670-106c-4fbc-b6a5-102896191190/kube-rbac-proxy/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.813125 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-htklg_af548670-106c-4fbc-b6a5-102896191190/manager/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.846474 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-j2drs_5949f2c9-f510-47f4-988c-cae06576536a/manager/0.log" Dec 10 10:33:50 crc kubenswrapper[4715]: I1210 10:33:50.884481 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-j2drs_5949f2c9-f510-47f4-988c-cae06576536a/kube-rbac-proxy/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.032170 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-scbxh_6c14bbb7-a3cf-4654-ba2c-2d993022fa72/kube-rbac-proxy/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.196391 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-scbxh_6c14bbb7-a3cf-4654-ba2c-2d993022fa72/manager/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.205232 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jdcr5_c396c828-63cf-4cd7-a050-c359ce8e1c8b/kube-rbac-proxy/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.292708 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jdcr5_c396c828-63cf-4cd7-a050-c359ce8e1c8b/manager/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.332651 4715 generic.go:334] "Generic (PLEG): container finished" podID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerID="3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd" exitCode=0 Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.332690 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4fdt" event={"ID":"c945121a-84b7-4a85-ae5c-ad652ceba703","Type":"ContainerDied","Data":"3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd"} Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.335288 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.378292 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-d875q_06228f2c-8f0c-493a-850f-b1e5fd41a23b/kube-rbac-proxy/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.448681 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-d875q_06228f2c-8f0c-493a-850f-b1e5fd41a23b/manager/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.566671 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-mxxcb_31619ce3-146d-4efb-898f-75bd1e007de9/kube-rbac-proxy/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.736331 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-h9rpn_c3397611-a5cb-455c-8c5e-dd24ab61545a/kube-rbac-proxy/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.779031 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-h9rpn_c3397611-a5cb-455c-8c5e-dd24ab61545a/manager/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.873528 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-mxxcb_31619ce3-146d-4efb-898f-75bd1e007de9/manager/0.log" Dec 10 10:33:51 crc kubenswrapper[4715]: I1210 10:33:51.955338 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-88z54_75ebabe2-8b57-4eee-9b3b-b238fafd94b7/kube-rbac-proxy/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.104361 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-88z54_75ebabe2-8b57-4eee-9b3b-b238fafd94b7/manager/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.107177 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-p69tc_d80992c0-9c12-4266-b705-6eaed7e46d1e/kube-rbac-proxy/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.194108 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-p69tc_d80992c0-9c12-4266-b705-6eaed7e46d1e/manager/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.298777 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-tjqbt_78ba49c9-f451-4982-8d88-6bf0444c8b60/kube-rbac-proxy/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.346352 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4fdt" event={"ID":"c945121a-84b7-4a85-ae5c-ad652ceba703","Type":"ContainerStarted","Data":"7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4"} Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.375490 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-tjqbt_78ba49c9-f451-4982-8d88-6bf0444c8b60/manager/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.527266 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-z25gz_20025d92-9314-436f-a05e-4bfee3d3add0/kube-rbac-proxy/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.581109 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-z25gz_20025d92-9314-436f-a05e-4bfee3d3add0/manager/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.618866 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-f8zw7_50f2991b-7b13-4d84-8a4d-504d9b873023/kube-rbac-proxy/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.787964 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-f8zw7_50f2991b-7b13-4d84-8a4d-504d9b873023/manager/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.825474 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-gbrf4_91566e3d-fbed-450a-bc66-1bf34f98d57f/kube-rbac-proxy/0.log" Dec 10 10:33:52 crc kubenswrapper[4715]: I1210 10:33:52.845537 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-gbrf4_91566e3d-fbed-450a-bc66-1bf34f98d57f/manager/0.log" Dec 10 10:33:53 crc kubenswrapper[4715]: I1210 10:33:53.045411 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fphth6_d329770a-7d39-41d0-b4b1-a52e2e9ccda7/manager/0.log" Dec 10 10:33:53 crc kubenswrapper[4715]: I1210 10:33:53.260044 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fphth6_d329770a-7d39-41d0-b4b1-a52e2e9ccda7/kube-rbac-proxy/0.log" Dec 10 10:33:53 crc kubenswrapper[4715]: I1210 10:33:53.357729 4715 generic.go:334] "Generic (PLEG): container finished" podID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerID="7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4" exitCode=0 Dec 10 10:33:53 crc kubenswrapper[4715]: I1210 10:33:53.357871 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4fdt" event={"ID":"c945121a-84b7-4a85-ae5c-ad652ceba703","Type":"ContainerDied","Data":"7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4"} Dec 10 10:33:53 crc kubenswrapper[4715]: I1210 10:33:53.457721 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-crsjq_313f8227-63e5-435e-a2e9-0c08d67712b8/registry-server/0.log" Dec 10 10:33:53 crc kubenswrapper[4715]: I1210 10:33:53.592061 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7dcf9bf79d-7ftkw_5899529c-0209-4c8f-8b67-37c9d08bbcf8/operator/0.log" Dec 10 10:33:53 crc kubenswrapper[4715]: I1210 10:33:53.773737 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-66fzd_01645280-6af5-4f8d-b412-3d7e0efa9054/kube-rbac-proxy/0.log" Dec 10 10:33:53 crc kubenswrapper[4715]: I1210 10:33:53.907621 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-m75xm_629d31d3-4474-4beb-9c7b-2d890d2134e6/kube-rbac-proxy/0.log" Dec 10 10:33:53 crc kubenswrapper[4715]: I1210 10:33:53.932255 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-66fzd_01645280-6af5-4f8d-b412-3d7e0efa9054/manager/0.log" Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.067133 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-m75xm_629d31d3-4474-4beb-9c7b-2d890d2134e6/manager/0.log" Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.222132 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-69rd4_3dccc54c-41e2-4745-839c-df51ca4e1825/operator/0.log" Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.326254 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-ht6n4_67b3cab4-c74e-4861-875b-e38aac7ced71/kube-rbac-proxy/0.log" Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.373682 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4fdt" event={"ID":"c945121a-84b7-4a85-ae5c-ad652ceba703","Type":"ContainerStarted","Data":"ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3"} Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.396004 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t4fdt" podStartSLOduration=2.929537769 podStartE2EDuration="5.395978861s" podCreationTimestamp="2025-12-10 10:33:49 +0000 UTC" firstStartedPulling="2025-12-10 10:33:51.334970137 +0000 UTC m=+3594.078516378" lastFinishedPulling="2025-12-10 10:33:53.801411219 +0000 UTC m=+3596.544957470" observedRunningTime="2025-12-10 10:33:54.393007636 +0000 UTC m=+3597.136553907" watchObservedRunningTime="2025-12-10 10:33:54.395978861 +0000 UTC m=+3597.139525112" Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.524756 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-ht6n4_67b3cab4-c74e-4861-875b-e38aac7ced71/manager/0.log" Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.659167 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-2957k_63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd/kube-rbac-proxy/0.log" Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.688177 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fd7979c46-c59rw_8c747bc0-5cc3-4cd2-8283-bdb2f2742167/manager/0.log" Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.828884 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-2957k_63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd/manager/0.log" Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.938086 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wcvvl_cfe88ca5-7d1d-4fff-8404-8b192ae48b09/manager/0.log" Dec 10 10:33:54 crc kubenswrapper[4715]: I1210 10:33:54.982580 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wcvvl_cfe88ca5-7d1d-4fff-8404-8b192ae48b09/kube-rbac-proxy/0.log" Dec 10 10:33:55 crc kubenswrapper[4715]: I1210 10:33:55.071181 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-x98sw_d66b6939-c9a0-413b-b09a-5eb529657837/kube-rbac-proxy/0.log" Dec 10 10:33:55 crc kubenswrapper[4715]: I1210 10:33:55.197675 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-x98sw_d66b6939-c9a0-413b-b09a-5eb529657837/manager/0.log" Dec 10 10:33:59 crc kubenswrapper[4715]: I1210 10:33:59.685805 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:59 crc kubenswrapper[4715]: I1210 10:33:59.686333 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:33:59 crc kubenswrapper[4715]: I1210 10:33:59.739794 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:34:00 crc kubenswrapper[4715]: I1210 10:34:00.485044 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:34:00 crc kubenswrapper[4715]: I1210 10:34:00.549317 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4fdt"] Dec 10 10:34:02 crc kubenswrapper[4715]: I1210 10:34:02.439487 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t4fdt" podUID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerName="registry-server" containerID="cri-o://ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3" gracePeriod=2 Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.002732 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.050403 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hn6fl\" (UniqueName: \"kubernetes.io/projected/c945121a-84b7-4a85-ae5c-ad652ceba703-kube-api-access-hn6fl\") pod \"c945121a-84b7-4a85-ae5c-ad652ceba703\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.050491 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-utilities\") pod \"c945121a-84b7-4a85-ae5c-ad652ceba703\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.050559 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-catalog-content\") pod \"c945121a-84b7-4a85-ae5c-ad652ceba703\" (UID: \"c945121a-84b7-4a85-ae5c-ad652ceba703\") " Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.051317 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-utilities" (OuterVolumeSpecName: "utilities") pod "c945121a-84b7-4a85-ae5c-ad652ceba703" (UID: "c945121a-84b7-4a85-ae5c-ad652ceba703"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.068171 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c945121a-84b7-4a85-ae5c-ad652ceba703-kube-api-access-hn6fl" (OuterVolumeSpecName: "kube-api-access-hn6fl") pod "c945121a-84b7-4a85-ae5c-ad652ceba703" (UID: "c945121a-84b7-4a85-ae5c-ad652ceba703"). InnerVolumeSpecName "kube-api-access-hn6fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.072740 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c945121a-84b7-4a85-ae5c-ad652ceba703" (UID: "c945121a-84b7-4a85-ae5c-ad652ceba703"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.152489 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.153007 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hn6fl\" (UniqueName: \"kubernetes.io/projected/c945121a-84b7-4a85-ae5c-ad652ceba703-kube-api-access-hn6fl\") on node \"crc\" DevicePath \"\"" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.153123 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c945121a-84b7-4a85-ae5c-ad652ceba703-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.470013 4715 generic.go:334] "Generic (PLEG): container finished" podID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerID="ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3" exitCode=0 Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.470076 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4fdt" event={"ID":"c945121a-84b7-4a85-ae5c-ad652ceba703","Type":"ContainerDied","Data":"ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3"} Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.470118 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4fdt" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.470174 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4fdt" event={"ID":"c945121a-84b7-4a85-ae5c-ad652ceba703","Type":"ContainerDied","Data":"39b451f508a376232b8f084a26671360051275255b0bcd41f2ad1e880506ec1e"} Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.470199 4715 scope.go:117] "RemoveContainer" containerID="ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.500396 4715 scope.go:117] "RemoveContainer" containerID="7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.505882 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4fdt"] Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.514486 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4fdt"] Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.539565 4715 scope.go:117] "RemoveContainer" containerID="3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.575354 4715 scope.go:117] "RemoveContainer" containerID="ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3" Dec 10 10:34:03 crc kubenswrapper[4715]: E1210 10:34:03.578446 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3\": container with ID starting with ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3 not found: ID does not exist" containerID="ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.578496 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3"} err="failed to get container status \"ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3\": rpc error: code = NotFound desc = could not find container \"ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3\": container with ID starting with ff0e793d5c058efa12ff5194f6a44b9fa8b0c0d92e9721bf39ca2e7656c3c3a3 not found: ID does not exist" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.578527 4715 scope.go:117] "RemoveContainer" containerID="7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4" Dec 10 10:34:03 crc kubenswrapper[4715]: E1210 10:34:03.578789 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4\": container with ID starting with 7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4 not found: ID does not exist" containerID="7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.578821 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4"} err="failed to get container status \"7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4\": rpc error: code = NotFound desc = could not find container \"7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4\": container with ID starting with 7b7ba10d4bb2772da344327ae96bcc583f04fede4432c61c0888e1d16e7f92c4 not found: ID does not exist" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.578840 4715 scope.go:117] "RemoveContainer" containerID="3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd" Dec 10 10:34:03 crc kubenswrapper[4715]: E1210 10:34:03.579090 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd\": container with ID starting with 3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd not found: ID does not exist" containerID="3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.579110 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd"} err="failed to get container status \"3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd\": rpc error: code = NotFound desc = could not find container \"3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd\": container with ID starting with 3f596d2580fe190fa4d747be2b0cda343b13531d2c4648a7e12296f5281b1cdd not found: ID does not exist" Dec 10 10:34:03 crc kubenswrapper[4715]: I1210 10:34:03.616767 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c945121a-84b7-4a85-ae5c-ad652ceba703" path="/var/lib/kubelet/pods/c945121a-84b7-4a85-ae5c-ad652ceba703/volumes" Dec 10 10:34:13 crc kubenswrapper[4715]: I1210 10:34:13.659691 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-tnrn5_45bf2bc4-681f-46cf-aca3-93960c133428/control-plane-machine-set-operator/0.log" Dec 10 10:34:13 crc kubenswrapper[4715]: I1210 10:34:13.809748 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fd9v5_c9ce5159-4cdb-47c7-9193-8adde2d4e2d0/kube-rbac-proxy/0.log" Dec 10 10:34:13 crc kubenswrapper[4715]: I1210 10:34:13.835647 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fd9v5_c9ce5159-4cdb-47c7-9193-8adde2d4e2d0/machine-api-operator/0.log" Dec 10 10:34:26 crc kubenswrapper[4715]: I1210 10:34:26.309985 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-mmntd_7a38a68d-c3fe-40f5-be5d-4c75b919c4c4/cert-manager-controller/0.log" Dec 10 10:34:26 crc kubenswrapper[4715]: I1210 10:34:26.419616 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-cxn4s_189f8af2-11e6-4140-8104-fb4a66bce776/cert-manager-cainjector/0.log" Dec 10 10:34:26 crc kubenswrapper[4715]: I1210 10:34:26.540033 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-vckw6_7f4c1366-2e47-4899-a38f-55cbc26f0569/cert-manager-webhook/0.log" Dec 10 10:34:37 crc kubenswrapper[4715]: I1210 10:34:37.736784 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-9srkc_2993b76a-4f33-4a02-8845-e694f2f65179/nmstate-console-plugin/0.log" Dec 10 10:34:37 crc kubenswrapper[4715]: I1210 10:34:37.915187 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-hhxxc_11a0c637-5b67-4db5-8fd0-33c6fabc6103/nmstate-handler/0.log" Dec 10 10:34:37 crc kubenswrapper[4715]: I1210 10:34:37.961022 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lwx7z_adb59670-27dd-447f-8ccc-ca87956810ec/nmstate-metrics/0.log" Dec 10 10:34:37 crc kubenswrapper[4715]: I1210 10:34:37.998390 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lwx7z_adb59670-27dd-447f-8ccc-ca87956810ec/kube-rbac-proxy/0.log" Dec 10 10:34:38 crc kubenswrapper[4715]: I1210 10:34:38.169800 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-bztww_bb6d3ad9-978f-48ff-b72c-434fa2c5783e/nmstate-operator/0.log" Dec 10 10:34:38 crc kubenswrapper[4715]: I1210 10:34:38.203593 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-nsj47_bc0de16a-d1b8-41b0-ad21-6015ea156e9a/nmstate-webhook/0.log" Dec 10 10:34:52 crc kubenswrapper[4715]: I1210 10:34:52.532373 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-xsr9b_fd252c98-3f30-4b3f-aa36-4056f00035a2/kube-rbac-proxy/0.log" Dec 10 10:34:52 crc kubenswrapper[4715]: I1210 10:34:52.659685 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-xsr9b_fd252c98-3f30-4b3f-aa36-4056f00035a2/controller/0.log" Dec 10 10:34:52 crc kubenswrapper[4715]: I1210 10:34:52.921606 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-frr-files/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.135547 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-reloader/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.157610 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-frr-files/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.177588 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-metrics/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.178227 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-reloader/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.338559 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-frr-files/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.372367 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-reloader/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.383316 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-metrics/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.409236 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-metrics/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.561637 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-reloader/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.616682 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-frr-files/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.638702 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-metrics/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.679874 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/controller/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.850137 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/kube-rbac-proxy/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.864048 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/frr-metrics/0.log" Dec 10 10:34:53 crc kubenswrapper[4715]: I1210 10:34:53.996801 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/kube-rbac-proxy-frr/0.log" Dec 10 10:34:54 crc kubenswrapper[4715]: I1210 10:34:54.112782 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/reloader/0.log" Dec 10 10:34:54 crc kubenswrapper[4715]: I1210 10:34:54.209352 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-s5rwg_e7f61745-e2a2-4e9a-b2a1-b931599251a8/frr-k8s-webhook-server/0.log" Dec 10 10:34:54 crc kubenswrapper[4715]: I1210 10:34:54.395562 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-69456cd765-w9pmv_6b85b47f-761c-4064-8a3e-21824cb5eef7/manager/0.log" Dec 10 10:34:54 crc kubenswrapper[4715]: I1210 10:34:54.570447 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-79dd69b788-gwkcv_711fc7a7-527a-4e3b-b343-a279ce6d3b79/webhook-server/0.log" Dec 10 10:34:54 crc kubenswrapper[4715]: I1210 10:34:54.652491 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-g6jgv_6f9bf056-705e-4f97-a470-1bd98ae14e30/kube-rbac-proxy/0.log" Dec 10 10:34:55 crc kubenswrapper[4715]: I1210 10:34:55.226702 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-g6jgv_6f9bf056-705e-4f97-a470-1bd98ae14e30/speaker/0.log" Dec 10 10:34:55 crc kubenswrapper[4715]: I1210 10:34:55.246980 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/frr/0.log" Dec 10 10:35:06 crc kubenswrapper[4715]: I1210 10:35:06.436790 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/util/0.log" Dec 10 10:35:06 crc kubenswrapper[4715]: I1210 10:35:06.626361 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/pull/0.log" Dec 10 10:35:06 crc kubenswrapper[4715]: I1210 10:35:06.648055 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/pull/0.log" Dec 10 10:35:06 crc kubenswrapper[4715]: I1210 10:35:06.651280 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/util/0.log" Dec 10 10:35:06 crc kubenswrapper[4715]: I1210 10:35:06.799169 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/extract/0.log" Dec 10 10:35:06 crc kubenswrapper[4715]: I1210 10:35:06.799332 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/util/0.log" Dec 10 10:35:06 crc kubenswrapper[4715]: I1210 10:35:06.848001 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/pull/0.log" Dec 10 10:35:06 crc kubenswrapper[4715]: I1210 10:35:06.976614 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/util/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.172729 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/util/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.188643 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/pull/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.208963 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/pull/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.333943 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/util/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.363554 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/pull/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.386822 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/extract/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.518391 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-utilities/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.682453 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-content/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.692774 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-utilities/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.718676 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-content/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.893674 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-content/0.log" Dec 10 10:35:07 crc kubenswrapper[4715]: I1210 10:35:07.967262 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-utilities/0.log" Dec 10 10:35:08 crc kubenswrapper[4715]: I1210 10:35:08.007893 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/registry-server/0.log" Dec 10 10:35:08 crc kubenswrapper[4715]: I1210 10:35:08.139985 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-utilities/0.log" Dec 10 10:35:08 crc kubenswrapper[4715]: I1210 10:35:08.286314 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-utilities/0.log" Dec 10 10:35:08 crc kubenswrapper[4715]: I1210 10:35:08.300567 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-content/0.log" Dec 10 10:35:08 crc kubenswrapper[4715]: I1210 10:35:08.319594 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-content/0.log" Dec 10 10:35:08 crc kubenswrapper[4715]: I1210 10:35:08.465953 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-content/0.log" Dec 10 10:35:08 crc kubenswrapper[4715]: I1210 10:35:08.471363 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-utilities/0.log" Dec 10 10:35:08 crc kubenswrapper[4715]: I1210 10:35:08.704048 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-rjxvk_ed4c45ce-d179-4515-8166-9f1c6c7e0913/marketplace-operator/0.log" Dec 10 10:35:08 crc kubenswrapper[4715]: I1210 10:35:08.802256 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-utilities/0.log" Dec 10 10:35:08 crc kubenswrapper[4715]: I1210 10:35:08.954899 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-utilities/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.056458 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-content/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.079792 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/registry-server/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.085153 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-content/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.261685 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-content/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.279281 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-utilities/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.327883 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/registry-server/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.497289 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-utilities/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.626779 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-content/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.630109 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-utilities/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.655506 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-content/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.821805 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-utilities/0.log" Dec 10 10:35:09 crc kubenswrapper[4715]: I1210 10:35:09.822960 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-content/0.log" Dec 10 10:35:10 crc kubenswrapper[4715]: I1210 10:35:10.338826 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/registry-server/0.log" Dec 10 10:35:44 crc kubenswrapper[4715]: E1210 10:35:44.523635 4715 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.201:49104->38.102.83.201:39121: read tcp 38.102.83.201:49104->38.102.83.201:39121: read: connection reset by peer Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.650239 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p64rg"] Dec 10 10:36:08 crc kubenswrapper[4715]: E1210 10:36:08.651680 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerName="extract-utilities" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.651698 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerName="extract-utilities" Dec 10 10:36:08 crc kubenswrapper[4715]: E1210 10:36:08.651720 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerName="registry-server" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.651735 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerName="registry-server" Dec 10 10:36:08 crc kubenswrapper[4715]: E1210 10:36:08.651757 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerName="extract-content" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.651763 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerName="extract-content" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.652008 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c945121a-84b7-4a85-ae5c-ad652ceba703" containerName="registry-server" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.653902 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.674828 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p64rg"] Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.834427 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-utilities\") pod \"redhat-operators-p64rg\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.834591 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-catalog-content\") pod \"redhat-operators-p64rg\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.834678 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld7b8\" (UniqueName: \"kubernetes.io/projected/5b8989bb-0242-4fe5-8969-9335fd360cda-kube-api-access-ld7b8\") pod \"redhat-operators-p64rg\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.937061 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-utilities\") pod \"redhat-operators-p64rg\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.937223 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-catalog-content\") pod \"redhat-operators-p64rg\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.937296 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld7b8\" (UniqueName: \"kubernetes.io/projected/5b8989bb-0242-4fe5-8969-9335fd360cda-kube-api-access-ld7b8\") pod \"redhat-operators-p64rg\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.937940 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-utilities\") pod \"redhat-operators-p64rg\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.937962 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-catalog-content\") pod \"redhat-operators-p64rg\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.964811 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld7b8\" (UniqueName: \"kubernetes.io/projected/5b8989bb-0242-4fe5-8969-9335fd360cda-kube-api-access-ld7b8\") pod \"redhat-operators-p64rg\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:08 crc kubenswrapper[4715]: I1210 10:36:08.982727 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:09 crc kubenswrapper[4715]: I1210 10:36:09.533406 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p64rg"] Dec 10 10:36:10 crc kubenswrapper[4715]: I1210 10:36:10.172067 4715 generic.go:334] "Generic (PLEG): container finished" podID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerID="e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc" exitCode=0 Dec 10 10:36:10 crc kubenswrapper[4715]: I1210 10:36:10.172167 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p64rg" event={"ID":"5b8989bb-0242-4fe5-8969-9335fd360cda","Type":"ContainerDied","Data":"e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc"} Dec 10 10:36:10 crc kubenswrapper[4715]: I1210 10:36:10.172430 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p64rg" event={"ID":"5b8989bb-0242-4fe5-8969-9335fd360cda","Type":"ContainerStarted","Data":"5ca451e481fc36acccb59eb571abc8dccb917f1b5af9d7de1a1886e36fc04b45"} Dec 10 10:36:13 crc kubenswrapper[4715]: I1210 10:36:13.208499 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p64rg" event={"ID":"5b8989bb-0242-4fe5-8969-9335fd360cda","Type":"ContainerStarted","Data":"44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1"} Dec 10 10:36:15 crc kubenswrapper[4715]: I1210 10:36:15.239680 4715 generic.go:334] "Generic (PLEG): container finished" podID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerID="44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1" exitCode=0 Dec 10 10:36:15 crc kubenswrapper[4715]: I1210 10:36:15.239741 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p64rg" event={"ID":"5b8989bb-0242-4fe5-8969-9335fd360cda","Type":"ContainerDied","Data":"44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1"} Dec 10 10:36:16 crc kubenswrapper[4715]: I1210 10:36:16.254145 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p64rg" event={"ID":"5b8989bb-0242-4fe5-8969-9335fd360cda","Type":"ContainerStarted","Data":"e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11"} Dec 10 10:36:16 crc kubenswrapper[4715]: I1210 10:36:16.277356 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p64rg" podStartSLOduration=2.69118381 podStartE2EDuration="8.277333172s" podCreationTimestamp="2025-12-10 10:36:08 +0000 UTC" firstStartedPulling="2025-12-10 10:36:10.174340149 +0000 UTC m=+3732.917886400" lastFinishedPulling="2025-12-10 10:36:15.760489511 +0000 UTC m=+3738.504035762" observedRunningTime="2025-12-10 10:36:16.271035563 +0000 UTC m=+3739.014581844" watchObservedRunningTime="2025-12-10 10:36:16.277333172 +0000 UTC m=+3739.020879423" Dec 10 10:36:17 crc kubenswrapper[4715]: I1210 10:36:17.714742 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:36:17 crc kubenswrapper[4715]: I1210 10:36:17.715337 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:36:18 crc kubenswrapper[4715]: I1210 10:36:18.984060 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:18 crc kubenswrapper[4715]: I1210 10:36:18.984116 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:20 crc kubenswrapper[4715]: I1210 10:36:20.039609 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p64rg" podUID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerName="registry-server" probeResult="failure" output=< Dec 10 10:36:20 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 10 10:36:20 crc kubenswrapper[4715]: > Dec 10 10:36:29 crc kubenswrapper[4715]: I1210 10:36:29.061115 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:29 crc kubenswrapper[4715]: I1210 10:36:29.113492 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:30 crc kubenswrapper[4715]: I1210 10:36:30.294976 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p64rg"] Dec 10 10:36:30 crc kubenswrapper[4715]: I1210 10:36:30.426771 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p64rg" podUID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerName="registry-server" containerID="cri-o://e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11" gracePeriod=2 Dec 10 10:36:30 crc kubenswrapper[4715]: I1210 10:36:30.969131 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.092469 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld7b8\" (UniqueName: \"kubernetes.io/projected/5b8989bb-0242-4fe5-8969-9335fd360cda-kube-api-access-ld7b8\") pod \"5b8989bb-0242-4fe5-8969-9335fd360cda\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.092562 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-catalog-content\") pod \"5b8989bb-0242-4fe5-8969-9335fd360cda\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.092616 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-utilities\") pod \"5b8989bb-0242-4fe5-8969-9335fd360cda\" (UID: \"5b8989bb-0242-4fe5-8969-9335fd360cda\") " Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.093811 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-utilities" (OuterVolumeSpecName: "utilities") pod "5b8989bb-0242-4fe5-8969-9335fd360cda" (UID: "5b8989bb-0242-4fe5-8969-9335fd360cda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.100026 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b8989bb-0242-4fe5-8969-9335fd360cda-kube-api-access-ld7b8" (OuterVolumeSpecName: "kube-api-access-ld7b8") pod "5b8989bb-0242-4fe5-8969-9335fd360cda" (UID: "5b8989bb-0242-4fe5-8969-9335fd360cda"). InnerVolumeSpecName "kube-api-access-ld7b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.195278 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.195318 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld7b8\" (UniqueName: \"kubernetes.io/projected/5b8989bb-0242-4fe5-8969-9335fd360cda-kube-api-access-ld7b8\") on node \"crc\" DevicePath \"\"" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.210679 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b8989bb-0242-4fe5-8969-9335fd360cda" (UID: "5b8989bb-0242-4fe5-8969-9335fd360cda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.299369 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8989bb-0242-4fe5-8969-9335fd360cda-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.442055 4715 generic.go:334] "Generic (PLEG): container finished" podID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerID="e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11" exitCode=0 Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.442111 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p64rg" event={"ID":"5b8989bb-0242-4fe5-8969-9335fd360cda","Type":"ContainerDied","Data":"e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11"} Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.442143 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p64rg" event={"ID":"5b8989bb-0242-4fe5-8969-9335fd360cda","Type":"ContainerDied","Data":"5ca451e481fc36acccb59eb571abc8dccb917f1b5af9d7de1a1886e36fc04b45"} Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.442155 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p64rg" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.442161 4715 scope.go:117] "RemoveContainer" containerID="e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.491319 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p64rg"] Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.495645 4715 scope.go:117] "RemoveContainer" containerID="44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.505754 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p64rg"] Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.525866 4715 scope.go:117] "RemoveContainer" containerID="e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.587617 4715 scope.go:117] "RemoveContainer" containerID="e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11" Dec 10 10:36:31 crc kubenswrapper[4715]: E1210 10:36:31.588193 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11\": container with ID starting with e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11 not found: ID does not exist" containerID="e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.588257 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11"} err="failed to get container status \"e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11\": rpc error: code = NotFound desc = could not find container \"e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11\": container with ID starting with e7573c1c626b38bb55a30e08f11284e16694dc161a2fd5db4b8c99e7e1e5eb11 not found: ID does not exist" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.588293 4715 scope.go:117] "RemoveContainer" containerID="44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1" Dec 10 10:36:31 crc kubenswrapper[4715]: E1210 10:36:31.588875 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1\": container with ID starting with 44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1 not found: ID does not exist" containerID="44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.588927 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1"} err="failed to get container status \"44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1\": rpc error: code = NotFound desc = could not find container \"44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1\": container with ID starting with 44d5c603901015ec42ff7d37db853c2f39498d30edd53109b2dd4ebc6c9d17a1 not found: ID does not exist" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.588951 4715 scope.go:117] "RemoveContainer" containerID="e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc" Dec 10 10:36:31 crc kubenswrapper[4715]: E1210 10:36:31.589342 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc\": container with ID starting with e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc not found: ID does not exist" containerID="e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.589382 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc"} err="failed to get container status \"e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc\": rpc error: code = NotFound desc = could not find container \"e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc\": container with ID starting with e0241deaf36d690dc6c3f092f50b3567d20b048507cb9af0160098f0208f83cc not found: ID does not exist" Dec 10 10:36:31 crc kubenswrapper[4715]: I1210 10:36:31.618689 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b8989bb-0242-4fe5-8969-9335fd360cda" path="/var/lib/kubelet/pods/5b8989bb-0242-4fe5-8969-9335fd360cda/volumes" Dec 10 10:36:47 crc kubenswrapper[4715]: I1210 10:36:47.714185 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:36:47 crc kubenswrapper[4715]: I1210 10:36:47.714613 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:36:58 crc kubenswrapper[4715]: I1210 10:36:58.740661 4715 generic.go:334] "Generic (PLEG): container finished" podID="6c19a077-d0e9-46ab-8c3b-ab9406292b85" containerID="fcd7ff9db8716ad243c700648e501f79292c69e793f8b9df09334f8bee319e9d" exitCode=0 Dec 10 10:36:58 crc kubenswrapper[4715]: I1210 10:36:58.740770 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k97pd/must-gather-2j6sm" event={"ID":"6c19a077-d0e9-46ab-8c3b-ab9406292b85","Type":"ContainerDied","Data":"fcd7ff9db8716ad243c700648e501f79292c69e793f8b9df09334f8bee319e9d"} Dec 10 10:36:58 crc kubenswrapper[4715]: I1210 10:36:58.742096 4715 scope.go:117] "RemoveContainer" containerID="fcd7ff9db8716ad243c700648e501f79292c69e793f8b9df09334f8bee319e9d" Dec 10 10:36:59 crc kubenswrapper[4715]: I1210 10:36:59.058902 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-k97pd_must-gather-2j6sm_6c19a077-d0e9-46ab-8c3b-ab9406292b85/gather/0.log" Dec 10 10:37:06 crc kubenswrapper[4715]: I1210 10:37:06.613122 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-k97pd/must-gather-2j6sm"] Dec 10 10:37:06 crc kubenswrapper[4715]: I1210 10:37:06.615186 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-k97pd/must-gather-2j6sm" podUID="6c19a077-d0e9-46ab-8c3b-ab9406292b85" containerName="copy" containerID="cri-o://6ff4a052eab48f4016cf6b4e1a1cfcc20f9c15df35234d31e09c297b37fc2f2d" gracePeriod=2 Dec 10 10:37:06 crc kubenswrapper[4715]: I1210 10:37:06.621718 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-k97pd/must-gather-2j6sm"] Dec 10 10:37:06 crc kubenswrapper[4715]: I1210 10:37:06.824295 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-k97pd_must-gather-2j6sm_6c19a077-d0e9-46ab-8c3b-ab9406292b85/copy/0.log" Dec 10 10:37:06 crc kubenswrapper[4715]: I1210 10:37:06.824993 4715 generic.go:334] "Generic (PLEG): container finished" podID="6c19a077-d0e9-46ab-8c3b-ab9406292b85" containerID="6ff4a052eab48f4016cf6b4e1a1cfcc20f9c15df35234d31e09c297b37fc2f2d" exitCode=143 Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.108757 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-k97pd_must-gather-2j6sm_6c19a077-d0e9-46ab-8c3b-ab9406292b85/copy/0.log" Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.109846 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/must-gather-2j6sm" Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.183011 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9fsn\" (UniqueName: \"kubernetes.io/projected/6c19a077-d0e9-46ab-8c3b-ab9406292b85-kube-api-access-x9fsn\") pod \"6c19a077-d0e9-46ab-8c3b-ab9406292b85\" (UID: \"6c19a077-d0e9-46ab-8c3b-ab9406292b85\") " Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.183197 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6c19a077-d0e9-46ab-8c3b-ab9406292b85-must-gather-output\") pod \"6c19a077-d0e9-46ab-8c3b-ab9406292b85\" (UID: \"6c19a077-d0e9-46ab-8c3b-ab9406292b85\") " Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.188676 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c19a077-d0e9-46ab-8c3b-ab9406292b85-kube-api-access-x9fsn" (OuterVolumeSpecName: "kube-api-access-x9fsn") pod "6c19a077-d0e9-46ab-8c3b-ab9406292b85" (UID: "6c19a077-d0e9-46ab-8c3b-ab9406292b85"). InnerVolumeSpecName "kube-api-access-x9fsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.285901 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9fsn\" (UniqueName: \"kubernetes.io/projected/6c19a077-d0e9-46ab-8c3b-ab9406292b85-kube-api-access-x9fsn\") on node \"crc\" DevicePath \"\"" Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.331592 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c19a077-d0e9-46ab-8c3b-ab9406292b85-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6c19a077-d0e9-46ab-8c3b-ab9406292b85" (UID: "6c19a077-d0e9-46ab-8c3b-ab9406292b85"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.387377 4715 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6c19a077-d0e9-46ab-8c3b-ab9406292b85-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.616217 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c19a077-d0e9-46ab-8c3b-ab9406292b85" path="/var/lib/kubelet/pods/6c19a077-d0e9-46ab-8c3b-ab9406292b85/volumes" Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.875302 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-k97pd_must-gather-2j6sm_6c19a077-d0e9-46ab-8c3b-ab9406292b85/copy/0.log" Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.876456 4715 scope.go:117] "RemoveContainer" containerID="6ff4a052eab48f4016cf6b4e1a1cfcc20f9c15df35234d31e09c297b37fc2f2d" Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.876636 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k97pd/must-gather-2j6sm" Dec 10 10:37:07 crc kubenswrapper[4715]: I1210 10:37:07.906058 4715 scope.go:117] "RemoveContainer" containerID="fcd7ff9db8716ad243c700648e501f79292c69e793f8b9df09334f8bee319e9d" Dec 10 10:37:17 crc kubenswrapper[4715]: I1210 10:37:17.714024 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:37:17 crc kubenswrapper[4715]: I1210 10:37:17.714598 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:37:17 crc kubenswrapper[4715]: I1210 10:37:17.714647 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 10:37:17 crc kubenswrapper[4715]: I1210 10:37:17.715464 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 10:37:17 crc kubenswrapper[4715]: I1210 10:37:17.715521 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" gracePeriod=600 Dec 10 10:37:17 crc kubenswrapper[4715]: E1210 10:37:17.922700 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:37:17 crc kubenswrapper[4715]: I1210 10:37:17.970461 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" exitCode=0 Dec 10 10:37:17 crc kubenswrapper[4715]: I1210 10:37:17.970529 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5"} Dec 10 10:37:17 crc kubenswrapper[4715]: I1210 10:37:17.970591 4715 scope.go:117] "RemoveContainer" containerID="62c20bae2d69f89d01d1894e46f3b38c45cea4c280173ffb7123e37e92db2bf8" Dec 10 10:37:17 crc kubenswrapper[4715]: I1210 10:37:17.971363 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:37:17 crc kubenswrapper[4715]: E1210 10:37:17.972062 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:37:28 crc kubenswrapper[4715]: I1210 10:37:28.605153 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:37:28 crc kubenswrapper[4715]: E1210 10:37:28.606111 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:37:41 crc kubenswrapper[4715]: I1210 10:37:41.604712 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:37:41 crc kubenswrapper[4715]: E1210 10:37:41.605618 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:37:54 crc kubenswrapper[4715]: I1210 10:37:54.605958 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:37:54 crc kubenswrapper[4715]: E1210 10:37:54.607167 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:38:05 crc kubenswrapper[4715]: I1210 10:38:05.606039 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:38:05 crc kubenswrapper[4715]: E1210 10:38:05.606822 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:38:18 crc kubenswrapper[4715]: I1210 10:38:18.604910 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:38:18 crc kubenswrapper[4715]: E1210 10:38:18.605726 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:38:29 crc kubenswrapper[4715]: I1210 10:38:29.606161 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:38:29 crc kubenswrapper[4715]: E1210 10:38:29.606986 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:38:33 crc kubenswrapper[4715]: I1210 10:38:33.196357 4715 scope.go:117] "RemoveContainer" containerID="6ba138c9fd5e85c1854ebd672b13f91bc7455abb9ad7594bba674ffa9cf81015" Dec 10 10:38:41 crc kubenswrapper[4715]: I1210 10:38:41.605068 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:38:41 crc kubenswrapper[4715]: E1210 10:38:41.605788 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:38:52 crc kubenswrapper[4715]: I1210 10:38:52.604970 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:38:52 crc kubenswrapper[4715]: E1210 10:38:52.605761 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:39:03 crc kubenswrapper[4715]: I1210 10:39:03.604945 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:39:03 crc kubenswrapper[4715]: E1210 10:39:03.605774 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:39:15 crc kubenswrapper[4715]: I1210 10:39:15.605464 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:39:15 crc kubenswrapper[4715]: E1210 10:39:15.606169 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.225557 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7lbsk"] Dec 10 10:39:22 crc kubenswrapper[4715]: E1210 10:39:22.226471 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerName="extract-content" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.226487 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerName="extract-content" Dec 10 10:39:22 crc kubenswrapper[4715]: E1210 10:39:22.226499 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c19a077-d0e9-46ab-8c3b-ab9406292b85" containerName="gather" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.226506 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c19a077-d0e9-46ab-8c3b-ab9406292b85" containerName="gather" Dec 10 10:39:22 crc kubenswrapper[4715]: E1210 10:39:22.226542 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerName="registry-server" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.226550 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerName="registry-server" Dec 10 10:39:22 crc kubenswrapper[4715]: E1210 10:39:22.226563 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerName="extract-utilities" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.226571 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerName="extract-utilities" Dec 10 10:39:22 crc kubenswrapper[4715]: E1210 10:39:22.226591 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c19a077-d0e9-46ab-8c3b-ab9406292b85" containerName="copy" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.226598 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c19a077-d0e9-46ab-8c3b-ab9406292b85" containerName="copy" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.226825 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b8989bb-0242-4fe5-8969-9335fd360cda" containerName="registry-server" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.226853 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c19a077-d0e9-46ab-8c3b-ab9406292b85" containerName="gather" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.226874 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c19a077-d0e9-46ab-8c3b-ab9406292b85" containerName="copy" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.228280 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.242234 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7lbsk"] Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.358067 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9csxd\" (UniqueName: \"kubernetes.io/projected/5c63e0b0-b997-4517-b86f-f1991f564b45-kube-api-access-9csxd\") pod \"certified-operators-7lbsk\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.358142 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-catalog-content\") pod \"certified-operators-7lbsk\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.358179 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-utilities\") pod \"certified-operators-7lbsk\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.459883 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9csxd\" (UniqueName: \"kubernetes.io/projected/5c63e0b0-b997-4517-b86f-f1991f564b45-kube-api-access-9csxd\") pod \"certified-operators-7lbsk\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.460006 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-catalog-content\") pod \"certified-operators-7lbsk\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.460059 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-utilities\") pod \"certified-operators-7lbsk\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.460575 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-catalog-content\") pod \"certified-operators-7lbsk\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.460654 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-utilities\") pod \"certified-operators-7lbsk\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.490790 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9csxd\" (UniqueName: \"kubernetes.io/projected/5c63e0b0-b997-4517-b86f-f1991f564b45-kube-api-access-9csxd\") pod \"certified-operators-7lbsk\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:22 crc kubenswrapper[4715]: I1210 10:39:22.548554 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:23 crc kubenswrapper[4715]: I1210 10:39:23.064966 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7lbsk"] Dec 10 10:39:23 crc kubenswrapper[4715]: I1210 10:39:23.188803 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lbsk" event={"ID":"5c63e0b0-b997-4517-b86f-f1991f564b45","Type":"ContainerStarted","Data":"28cfcf7246886566d6f85b419beeb801280f22c48b6fc3b88b48f6be8d467793"} Dec 10 10:39:24 crc kubenswrapper[4715]: I1210 10:39:24.199670 4715 generic.go:334] "Generic (PLEG): container finished" podID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerID="b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e" exitCode=0 Dec 10 10:39:24 crc kubenswrapper[4715]: I1210 10:39:24.199715 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lbsk" event={"ID":"5c63e0b0-b997-4517-b86f-f1991f564b45","Type":"ContainerDied","Data":"b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e"} Dec 10 10:39:24 crc kubenswrapper[4715]: I1210 10:39:24.203214 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 10:39:25 crc kubenswrapper[4715]: I1210 10:39:25.210646 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lbsk" event={"ID":"5c63e0b0-b997-4517-b86f-f1991f564b45","Type":"ContainerStarted","Data":"2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc"} Dec 10 10:39:26 crc kubenswrapper[4715]: I1210 10:39:26.224386 4715 generic.go:334] "Generic (PLEG): container finished" podID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerID="2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc" exitCode=0 Dec 10 10:39:26 crc kubenswrapper[4715]: I1210 10:39:26.224454 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lbsk" event={"ID":"5c63e0b0-b997-4517-b86f-f1991f564b45","Type":"ContainerDied","Data":"2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc"} Dec 10 10:39:28 crc kubenswrapper[4715]: I1210 10:39:28.250888 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lbsk" event={"ID":"5c63e0b0-b997-4517-b86f-f1991f564b45","Type":"ContainerStarted","Data":"8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101"} Dec 10 10:39:28 crc kubenswrapper[4715]: I1210 10:39:28.281537 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7lbsk" podStartSLOduration=2.7688367080000003 podStartE2EDuration="6.281511377s" podCreationTimestamp="2025-12-10 10:39:22 +0000 UTC" firstStartedPulling="2025-12-10 10:39:24.202840459 +0000 UTC m=+3926.946386710" lastFinishedPulling="2025-12-10 10:39:27.715515118 +0000 UTC m=+3930.459061379" observedRunningTime="2025-12-10 10:39:28.273342964 +0000 UTC m=+3931.016889225" watchObservedRunningTime="2025-12-10 10:39:28.281511377 +0000 UTC m=+3931.025057628" Dec 10 10:39:30 crc kubenswrapper[4715]: I1210 10:39:30.604745 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:39:30 crc kubenswrapper[4715]: E1210 10:39:30.605579 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:39:32 crc kubenswrapper[4715]: I1210 10:39:32.548806 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:32 crc kubenswrapper[4715]: I1210 10:39:32.549304 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:32 crc kubenswrapper[4715]: I1210 10:39:32.597041 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:33 crc kubenswrapper[4715]: I1210 10:39:33.357491 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:33 crc kubenswrapper[4715]: I1210 10:39:33.420711 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7lbsk"] Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.313505 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7lbsk" podUID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerName="registry-server" containerID="cri-o://8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101" gracePeriod=2 Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.731149 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.838461 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-utilities\") pod \"5c63e0b0-b997-4517-b86f-f1991f564b45\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.838511 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-catalog-content\") pod \"5c63e0b0-b997-4517-b86f-f1991f564b45\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.838653 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9csxd\" (UniqueName: \"kubernetes.io/projected/5c63e0b0-b997-4517-b86f-f1991f564b45-kube-api-access-9csxd\") pod \"5c63e0b0-b997-4517-b86f-f1991f564b45\" (UID: \"5c63e0b0-b997-4517-b86f-f1991f564b45\") " Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.839595 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-utilities" (OuterVolumeSpecName: "utilities") pod "5c63e0b0-b997-4517-b86f-f1991f564b45" (UID: "5c63e0b0-b997-4517-b86f-f1991f564b45"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.844434 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c63e0b0-b997-4517-b86f-f1991f564b45-kube-api-access-9csxd" (OuterVolumeSpecName: "kube-api-access-9csxd") pod "5c63e0b0-b997-4517-b86f-f1991f564b45" (UID: "5c63e0b0-b997-4517-b86f-f1991f564b45"). InnerVolumeSpecName "kube-api-access-9csxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.885986 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c63e0b0-b997-4517-b86f-f1991f564b45" (UID: "5c63e0b0-b997-4517-b86f-f1991f564b45"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.940942 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9csxd\" (UniqueName: \"kubernetes.io/projected/5c63e0b0-b997-4517-b86f-f1991f564b45-kube-api-access-9csxd\") on node \"crc\" DevicePath \"\"" Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.940985 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:39:35 crc kubenswrapper[4715]: I1210 10:39:35.940998 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c63e0b0-b997-4517-b86f-f1991f564b45-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.324171 4715 generic.go:334] "Generic (PLEG): container finished" podID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerID="8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101" exitCode=0 Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.324217 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lbsk" event={"ID":"5c63e0b0-b997-4517-b86f-f1991f564b45","Type":"ContainerDied","Data":"8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101"} Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.324274 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lbsk" event={"ID":"5c63e0b0-b997-4517-b86f-f1991f564b45","Type":"ContainerDied","Data":"28cfcf7246886566d6f85b419beeb801280f22c48b6fc3b88b48f6be8d467793"} Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.324297 4715 scope.go:117] "RemoveContainer" containerID="8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101" Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.324232 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7lbsk" Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.343849 4715 scope.go:117] "RemoveContainer" containerID="2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc" Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.361795 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7lbsk"] Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.372857 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7lbsk"] Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.389732 4715 scope.go:117] "RemoveContainer" containerID="b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e" Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.433134 4715 scope.go:117] "RemoveContainer" containerID="8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101" Dec 10 10:39:36 crc kubenswrapper[4715]: E1210 10:39:36.433654 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101\": container with ID starting with 8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101 not found: ID does not exist" containerID="8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101" Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.433694 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101"} err="failed to get container status \"8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101\": rpc error: code = NotFound desc = could not find container \"8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101\": container with ID starting with 8b51ffc2c67a652d04524f55b97f462d754079473542a8ed868b693b637bd101 not found: ID does not exist" Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.433723 4715 scope.go:117] "RemoveContainer" containerID="2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc" Dec 10 10:39:36 crc kubenswrapper[4715]: E1210 10:39:36.434025 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc\": container with ID starting with 2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc not found: ID does not exist" containerID="2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc" Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.434048 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc"} err="failed to get container status \"2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc\": rpc error: code = NotFound desc = could not find container \"2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc\": container with ID starting with 2bd6e24033c9388aa0f274c72502bf1f9bb3d9ddf0645d1cd775c25138a554fc not found: ID does not exist" Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.434063 4715 scope.go:117] "RemoveContainer" containerID="b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e" Dec 10 10:39:36 crc kubenswrapper[4715]: E1210 10:39:36.434529 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e\": container with ID starting with b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e not found: ID does not exist" containerID="b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e" Dec 10 10:39:36 crc kubenswrapper[4715]: I1210 10:39:36.434576 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e"} err="failed to get container status \"b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e\": rpc error: code = NotFound desc = could not find container \"b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e\": container with ID starting with b6eb2b599599b5a0eb5b5270d7f213c6d0bba1eec29eb423724609c9ae39dd9e not found: ID does not exist" Dec 10 10:39:37 crc kubenswrapper[4715]: I1210 10:39:37.618424 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c63e0b0-b997-4517-b86f-f1991f564b45" path="/var/lib/kubelet/pods/5c63e0b0-b997-4517-b86f-f1991f564b45/volumes" Dec 10 10:39:45 crc kubenswrapper[4715]: I1210 10:39:45.604708 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:39:45 crc kubenswrapper[4715]: E1210 10:39:45.605491 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:39:58 crc kubenswrapper[4715]: I1210 10:39:58.605050 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:39:58 crc kubenswrapper[4715]: E1210 10:39:58.606006 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.027625 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nr8ss/must-gather-4spj6"] Dec 10 10:40:02 crc kubenswrapper[4715]: E1210 10:40:02.028293 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerName="registry-server" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.028306 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerName="registry-server" Dec 10 10:40:02 crc kubenswrapper[4715]: E1210 10:40:02.028326 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerName="extract-content" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.028332 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerName="extract-content" Dec 10 10:40:02 crc kubenswrapper[4715]: E1210 10:40:02.028343 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerName="extract-utilities" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.028349 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerName="extract-utilities" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.028561 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c63e0b0-b997-4517-b86f-f1991f564b45" containerName="registry-server" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.030183 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/must-gather-4spj6" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.032858 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-nr8ss"/"default-dockercfg-2kh64" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.033572 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-nr8ss"/"openshift-service-ca.crt" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.033741 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-nr8ss"/"kube-root-ca.crt" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.038103 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nr8ss/must-gather-4spj6"] Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.132078 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94pjt\" (UniqueName: \"kubernetes.io/projected/8e73665f-3e61-4eda-8069-e05d39a614fc-kube-api-access-94pjt\") pod \"must-gather-4spj6\" (UID: \"8e73665f-3e61-4eda-8069-e05d39a614fc\") " pod="openshift-must-gather-nr8ss/must-gather-4spj6" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.132216 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8e73665f-3e61-4eda-8069-e05d39a614fc-must-gather-output\") pod \"must-gather-4spj6\" (UID: \"8e73665f-3e61-4eda-8069-e05d39a614fc\") " pod="openshift-must-gather-nr8ss/must-gather-4spj6" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.234539 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94pjt\" (UniqueName: \"kubernetes.io/projected/8e73665f-3e61-4eda-8069-e05d39a614fc-kube-api-access-94pjt\") pod \"must-gather-4spj6\" (UID: \"8e73665f-3e61-4eda-8069-e05d39a614fc\") " pod="openshift-must-gather-nr8ss/must-gather-4spj6" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.234998 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8e73665f-3e61-4eda-8069-e05d39a614fc-must-gather-output\") pod \"must-gather-4spj6\" (UID: \"8e73665f-3e61-4eda-8069-e05d39a614fc\") " pod="openshift-must-gather-nr8ss/must-gather-4spj6" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.235571 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8e73665f-3e61-4eda-8069-e05d39a614fc-must-gather-output\") pod \"must-gather-4spj6\" (UID: \"8e73665f-3e61-4eda-8069-e05d39a614fc\") " pod="openshift-must-gather-nr8ss/must-gather-4spj6" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.276834 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94pjt\" (UniqueName: \"kubernetes.io/projected/8e73665f-3e61-4eda-8069-e05d39a614fc-kube-api-access-94pjt\") pod \"must-gather-4spj6\" (UID: \"8e73665f-3e61-4eda-8069-e05d39a614fc\") " pod="openshift-must-gather-nr8ss/must-gather-4spj6" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.359436 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/must-gather-4spj6" Dec 10 10:40:02 crc kubenswrapper[4715]: I1210 10:40:02.874897 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nr8ss/must-gather-4spj6"] Dec 10 10:40:03 crc kubenswrapper[4715]: I1210 10:40:03.598733 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/must-gather-4spj6" event={"ID":"8e73665f-3e61-4eda-8069-e05d39a614fc","Type":"ContainerStarted","Data":"f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25"} Dec 10 10:40:03 crc kubenswrapper[4715]: I1210 10:40:03.599224 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/must-gather-4spj6" event={"ID":"8e73665f-3e61-4eda-8069-e05d39a614fc","Type":"ContainerStarted","Data":"0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e"} Dec 10 10:40:03 crc kubenswrapper[4715]: I1210 10:40:03.599240 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/must-gather-4spj6" event={"ID":"8e73665f-3e61-4eda-8069-e05d39a614fc","Type":"ContainerStarted","Data":"72a14ed2e9b3d8924bd0410f87e49ef80c2f0a9152e7e74901d0c7b9337be9f8"} Dec 10 10:40:03 crc kubenswrapper[4715]: I1210 10:40:03.616890 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-nr8ss/must-gather-4spj6" podStartSLOduration=1.616869695 podStartE2EDuration="1.616869695s" podCreationTimestamp="2025-12-10 10:40:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 10:40:03.611999236 +0000 UTC m=+3966.355545517" watchObservedRunningTime="2025-12-10 10:40:03.616869695 +0000 UTC m=+3966.360415946" Dec 10 10:40:07 crc kubenswrapper[4715]: I1210 10:40:07.141320 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nr8ss/crc-debug-qfshm"] Dec 10 10:40:07 crc kubenswrapper[4715]: I1210 10:40:07.144359 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-qfshm" Dec 10 10:40:07 crc kubenswrapper[4715]: I1210 10:40:07.277701 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62tmx\" (UniqueName: \"kubernetes.io/projected/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-kube-api-access-62tmx\") pod \"crc-debug-qfshm\" (UID: \"bd43048d-e7ce-4ada-8c35-9dc44eb7a735\") " pod="openshift-must-gather-nr8ss/crc-debug-qfshm" Dec 10 10:40:07 crc kubenswrapper[4715]: I1210 10:40:07.277858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-host\") pod \"crc-debug-qfshm\" (UID: \"bd43048d-e7ce-4ada-8c35-9dc44eb7a735\") " pod="openshift-must-gather-nr8ss/crc-debug-qfshm" Dec 10 10:40:07 crc kubenswrapper[4715]: I1210 10:40:07.380156 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-host\") pod \"crc-debug-qfshm\" (UID: \"bd43048d-e7ce-4ada-8c35-9dc44eb7a735\") " pod="openshift-must-gather-nr8ss/crc-debug-qfshm" Dec 10 10:40:07 crc kubenswrapper[4715]: I1210 10:40:07.380312 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-host\") pod \"crc-debug-qfshm\" (UID: \"bd43048d-e7ce-4ada-8c35-9dc44eb7a735\") " pod="openshift-must-gather-nr8ss/crc-debug-qfshm" Dec 10 10:40:07 crc kubenswrapper[4715]: I1210 10:40:07.380376 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62tmx\" (UniqueName: \"kubernetes.io/projected/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-kube-api-access-62tmx\") pod \"crc-debug-qfshm\" (UID: \"bd43048d-e7ce-4ada-8c35-9dc44eb7a735\") " pod="openshift-must-gather-nr8ss/crc-debug-qfshm" Dec 10 10:40:07 crc kubenswrapper[4715]: I1210 10:40:07.402409 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62tmx\" (UniqueName: \"kubernetes.io/projected/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-kube-api-access-62tmx\") pod \"crc-debug-qfshm\" (UID: \"bd43048d-e7ce-4ada-8c35-9dc44eb7a735\") " pod="openshift-must-gather-nr8ss/crc-debug-qfshm" Dec 10 10:40:07 crc kubenswrapper[4715]: I1210 10:40:07.463024 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-qfshm" Dec 10 10:40:07 crc kubenswrapper[4715]: I1210 10:40:07.647655 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/crc-debug-qfshm" event={"ID":"bd43048d-e7ce-4ada-8c35-9dc44eb7a735","Type":"ContainerStarted","Data":"b1d0daf77810c2bef82ca3628d4572d36972e8768cc9f76a030ead4154663265"} Dec 10 10:40:08 crc kubenswrapper[4715]: I1210 10:40:08.659438 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/crc-debug-qfshm" event={"ID":"bd43048d-e7ce-4ada-8c35-9dc44eb7a735","Type":"ContainerStarted","Data":"85126296ad76da71449282306d240cbb394ef0846f0478ef4302df17505c59da"} Dec 10 10:40:08 crc kubenswrapper[4715]: I1210 10:40:08.678494 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-nr8ss/crc-debug-qfshm" podStartSLOduration=1.67847214 podStartE2EDuration="1.67847214s" podCreationTimestamp="2025-12-10 10:40:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 10:40:08.67811764 +0000 UTC m=+3971.421663891" watchObservedRunningTime="2025-12-10 10:40:08.67847214 +0000 UTC m=+3971.422018391" Dec 10 10:40:11 crc kubenswrapper[4715]: I1210 10:40:11.605773 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:40:11 crc kubenswrapper[4715]: E1210 10:40:11.606534 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:40:24 crc kubenswrapper[4715]: I1210 10:40:24.605266 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:40:24 crc kubenswrapper[4715]: E1210 10:40:24.606064 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:40:35 crc kubenswrapper[4715]: I1210 10:40:35.605741 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:40:35 crc kubenswrapper[4715]: E1210 10:40:35.606745 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:40:43 crc kubenswrapper[4715]: I1210 10:40:43.095170 4715 generic.go:334] "Generic (PLEG): container finished" podID="bd43048d-e7ce-4ada-8c35-9dc44eb7a735" containerID="85126296ad76da71449282306d240cbb394ef0846f0478ef4302df17505c59da" exitCode=0 Dec 10 10:40:43 crc kubenswrapper[4715]: I1210 10:40:43.095256 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/crc-debug-qfshm" event={"ID":"bd43048d-e7ce-4ada-8c35-9dc44eb7a735","Type":"ContainerDied","Data":"85126296ad76da71449282306d240cbb394ef0846f0478ef4302df17505c59da"} Dec 10 10:40:44 crc kubenswrapper[4715]: I1210 10:40:44.200029 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-qfshm" Dec 10 10:40:44 crc kubenswrapper[4715]: I1210 10:40:44.237888 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nr8ss/crc-debug-qfshm"] Dec 10 10:40:44 crc kubenswrapper[4715]: I1210 10:40:44.249738 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nr8ss/crc-debug-qfshm"] Dec 10 10:40:44 crc kubenswrapper[4715]: I1210 10:40:44.336760 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-host\") pod \"bd43048d-e7ce-4ada-8c35-9dc44eb7a735\" (UID: \"bd43048d-e7ce-4ada-8c35-9dc44eb7a735\") " Dec 10 10:40:44 crc kubenswrapper[4715]: I1210 10:40:44.336935 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-host" (OuterVolumeSpecName: "host") pod "bd43048d-e7ce-4ada-8c35-9dc44eb7a735" (UID: "bd43048d-e7ce-4ada-8c35-9dc44eb7a735"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 10:40:44 crc kubenswrapper[4715]: I1210 10:40:44.336985 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62tmx\" (UniqueName: \"kubernetes.io/projected/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-kube-api-access-62tmx\") pod \"bd43048d-e7ce-4ada-8c35-9dc44eb7a735\" (UID: \"bd43048d-e7ce-4ada-8c35-9dc44eb7a735\") " Dec 10 10:40:44 crc kubenswrapper[4715]: I1210 10:40:44.337678 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-host\") on node \"crc\" DevicePath \"\"" Dec 10 10:40:44 crc kubenswrapper[4715]: I1210 10:40:44.349153 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-kube-api-access-62tmx" (OuterVolumeSpecName: "kube-api-access-62tmx") pod "bd43048d-e7ce-4ada-8c35-9dc44eb7a735" (UID: "bd43048d-e7ce-4ada-8c35-9dc44eb7a735"). InnerVolumeSpecName "kube-api-access-62tmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:40:44 crc kubenswrapper[4715]: I1210 10:40:44.439352 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62tmx\" (UniqueName: \"kubernetes.io/projected/bd43048d-e7ce-4ada-8c35-9dc44eb7a735-kube-api-access-62tmx\") on node \"crc\" DevicePath \"\"" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.115051 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1d0daf77810c2bef82ca3628d4572d36972e8768cc9f76a030ead4154663265" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.115135 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-qfshm" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.415217 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nr8ss/crc-debug-p578f"] Dec 10 10:40:45 crc kubenswrapper[4715]: E1210 10:40:45.415583 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd43048d-e7ce-4ada-8c35-9dc44eb7a735" containerName="container-00" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.415595 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd43048d-e7ce-4ada-8c35-9dc44eb7a735" containerName="container-00" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.415851 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd43048d-e7ce-4ada-8c35-9dc44eb7a735" containerName="container-00" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.416731 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-p578f" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.477419 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-host\") pod \"crc-debug-p578f\" (UID: \"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e\") " pod="openshift-must-gather-nr8ss/crc-debug-p578f" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.478054 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9jqr\" (UniqueName: \"kubernetes.io/projected/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-kube-api-access-b9jqr\") pod \"crc-debug-p578f\" (UID: \"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e\") " pod="openshift-must-gather-nr8ss/crc-debug-p578f" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.579177 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9jqr\" (UniqueName: \"kubernetes.io/projected/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-kube-api-access-b9jqr\") pod \"crc-debug-p578f\" (UID: \"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e\") " pod="openshift-must-gather-nr8ss/crc-debug-p578f" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.579245 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-host\") pod \"crc-debug-p578f\" (UID: \"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e\") " pod="openshift-must-gather-nr8ss/crc-debug-p578f" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.579401 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-host\") pod \"crc-debug-p578f\" (UID: \"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e\") " pod="openshift-must-gather-nr8ss/crc-debug-p578f" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.615433 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd43048d-e7ce-4ada-8c35-9dc44eb7a735" path="/var/lib/kubelet/pods/bd43048d-e7ce-4ada-8c35-9dc44eb7a735/volumes" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.615594 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9jqr\" (UniqueName: \"kubernetes.io/projected/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-kube-api-access-b9jqr\") pod \"crc-debug-p578f\" (UID: \"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e\") " pod="openshift-must-gather-nr8ss/crc-debug-p578f" Dec 10 10:40:45 crc kubenswrapper[4715]: I1210 10:40:45.732520 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-p578f" Dec 10 10:40:46 crc kubenswrapper[4715]: I1210 10:40:46.124361 4715 generic.go:334] "Generic (PLEG): container finished" podID="9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e" containerID="7b5894c1728bf53a09d8fdf4cb846a1f372b205e2986b0f78117634ddea27067" exitCode=0 Dec 10 10:40:46 crc kubenswrapper[4715]: I1210 10:40:46.124418 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/crc-debug-p578f" event={"ID":"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e","Type":"ContainerDied","Data":"7b5894c1728bf53a09d8fdf4cb846a1f372b205e2986b0f78117634ddea27067"} Dec 10 10:40:46 crc kubenswrapper[4715]: I1210 10:40:46.124454 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/crc-debug-p578f" event={"ID":"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e","Type":"ContainerStarted","Data":"05dddbf2afcc2e00e23ab9460a106d0be38b7bc950923a4588c2e7d9b50aa482"} Dec 10 10:40:46 crc kubenswrapper[4715]: I1210 10:40:46.605557 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:40:46 crc kubenswrapper[4715]: E1210 10:40:46.606322 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:40:46 crc kubenswrapper[4715]: I1210 10:40:46.658463 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nr8ss/crc-debug-p578f"] Dec 10 10:40:46 crc kubenswrapper[4715]: I1210 10:40:46.668059 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nr8ss/crc-debug-p578f"] Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.234155 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-p578f" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.327855 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9jqr\" (UniqueName: \"kubernetes.io/projected/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-kube-api-access-b9jqr\") pod \"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e\" (UID: \"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e\") " Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.328020 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-host\") pod \"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e\" (UID: \"9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e\") " Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.328208 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-host" (OuterVolumeSpecName: "host") pod "9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e" (UID: "9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.328675 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-host\") on node \"crc\" DevicePath \"\"" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.337014 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-kube-api-access-b9jqr" (OuterVolumeSpecName: "kube-api-access-b9jqr") pod "9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e" (UID: "9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e"). InnerVolumeSpecName "kube-api-access-b9jqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.430290 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9jqr\" (UniqueName: \"kubernetes.io/projected/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e-kube-api-access-b9jqr\") on node \"crc\" DevicePath \"\"" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.623401 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e" path="/var/lib/kubelet/pods/9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e/volumes" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.834959 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nr8ss/crc-debug-w72xn"] Dec 10 10:40:47 crc kubenswrapper[4715]: E1210 10:40:47.835542 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e" containerName="container-00" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.835562 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e" containerName="container-00" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.835770 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b4106a7-6cd7-4fc9-80fe-de25e2cf0b5e" containerName="container-00" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.836429 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-w72xn" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.939119 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46wjr\" (UniqueName: \"kubernetes.io/projected/00d08c4e-1767-4a43-88a2-dc3e5df676f1-kube-api-access-46wjr\") pod \"crc-debug-w72xn\" (UID: \"00d08c4e-1767-4a43-88a2-dc3e5df676f1\") " pod="openshift-must-gather-nr8ss/crc-debug-w72xn" Dec 10 10:40:47 crc kubenswrapper[4715]: I1210 10:40:47.939298 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/00d08c4e-1767-4a43-88a2-dc3e5df676f1-host\") pod \"crc-debug-w72xn\" (UID: \"00d08c4e-1767-4a43-88a2-dc3e5df676f1\") " pod="openshift-must-gather-nr8ss/crc-debug-w72xn" Dec 10 10:40:48 crc kubenswrapper[4715]: I1210 10:40:48.041053 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/00d08c4e-1767-4a43-88a2-dc3e5df676f1-host\") pod \"crc-debug-w72xn\" (UID: \"00d08c4e-1767-4a43-88a2-dc3e5df676f1\") " pod="openshift-must-gather-nr8ss/crc-debug-w72xn" Dec 10 10:40:48 crc kubenswrapper[4715]: I1210 10:40:48.041221 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46wjr\" (UniqueName: \"kubernetes.io/projected/00d08c4e-1767-4a43-88a2-dc3e5df676f1-kube-api-access-46wjr\") pod \"crc-debug-w72xn\" (UID: \"00d08c4e-1767-4a43-88a2-dc3e5df676f1\") " pod="openshift-must-gather-nr8ss/crc-debug-w72xn" Dec 10 10:40:48 crc kubenswrapper[4715]: I1210 10:40:48.041686 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/00d08c4e-1767-4a43-88a2-dc3e5df676f1-host\") pod \"crc-debug-w72xn\" (UID: \"00d08c4e-1767-4a43-88a2-dc3e5df676f1\") " pod="openshift-must-gather-nr8ss/crc-debug-w72xn" Dec 10 10:40:48 crc kubenswrapper[4715]: I1210 10:40:48.080257 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46wjr\" (UniqueName: \"kubernetes.io/projected/00d08c4e-1767-4a43-88a2-dc3e5df676f1-kube-api-access-46wjr\") pod \"crc-debug-w72xn\" (UID: \"00d08c4e-1767-4a43-88a2-dc3e5df676f1\") " pod="openshift-must-gather-nr8ss/crc-debug-w72xn" Dec 10 10:40:48 crc kubenswrapper[4715]: I1210 10:40:48.142770 4715 scope.go:117] "RemoveContainer" containerID="7b5894c1728bf53a09d8fdf4cb846a1f372b205e2986b0f78117634ddea27067" Dec 10 10:40:48 crc kubenswrapper[4715]: I1210 10:40:48.142797 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-p578f" Dec 10 10:40:48 crc kubenswrapper[4715]: I1210 10:40:48.151992 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-w72xn" Dec 10 10:40:49 crc kubenswrapper[4715]: I1210 10:40:49.171016 4715 generic.go:334] "Generic (PLEG): container finished" podID="00d08c4e-1767-4a43-88a2-dc3e5df676f1" containerID="a6ab83a5729c6650b0aa88ec31d614839166d697b805833fcd37cc9dfd80bd56" exitCode=0 Dec 10 10:40:49 crc kubenswrapper[4715]: I1210 10:40:49.171103 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/crc-debug-w72xn" event={"ID":"00d08c4e-1767-4a43-88a2-dc3e5df676f1","Type":"ContainerDied","Data":"a6ab83a5729c6650b0aa88ec31d614839166d697b805833fcd37cc9dfd80bd56"} Dec 10 10:40:49 crc kubenswrapper[4715]: I1210 10:40:49.171427 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/crc-debug-w72xn" event={"ID":"00d08c4e-1767-4a43-88a2-dc3e5df676f1","Type":"ContainerStarted","Data":"120cd596d64cca248b1d7079a0a1c745732b62bb0398e777fe2f0fc646f79748"} Dec 10 10:40:49 crc kubenswrapper[4715]: I1210 10:40:49.217402 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nr8ss/crc-debug-w72xn"] Dec 10 10:40:49 crc kubenswrapper[4715]: I1210 10:40:49.232854 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nr8ss/crc-debug-w72xn"] Dec 10 10:40:50 crc kubenswrapper[4715]: I1210 10:40:50.294866 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-w72xn" Dec 10 10:40:50 crc kubenswrapper[4715]: I1210 10:40:50.384628 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/00d08c4e-1767-4a43-88a2-dc3e5df676f1-host\") pod \"00d08c4e-1767-4a43-88a2-dc3e5df676f1\" (UID: \"00d08c4e-1767-4a43-88a2-dc3e5df676f1\") " Dec 10 10:40:50 crc kubenswrapper[4715]: I1210 10:40:50.384698 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46wjr\" (UniqueName: \"kubernetes.io/projected/00d08c4e-1767-4a43-88a2-dc3e5df676f1-kube-api-access-46wjr\") pod \"00d08c4e-1767-4a43-88a2-dc3e5df676f1\" (UID: \"00d08c4e-1767-4a43-88a2-dc3e5df676f1\") " Dec 10 10:40:50 crc kubenswrapper[4715]: I1210 10:40:50.384753 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00d08c4e-1767-4a43-88a2-dc3e5df676f1-host" (OuterVolumeSpecName: "host") pod "00d08c4e-1767-4a43-88a2-dc3e5df676f1" (UID: "00d08c4e-1767-4a43-88a2-dc3e5df676f1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 10:40:50 crc kubenswrapper[4715]: I1210 10:40:50.385283 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/00d08c4e-1767-4a43-88a2-dc3e5df676f1-host\") on node \"crc\" DevicePath \"\"" Dec 10 10:40:50 crc kubenswrapper[4715]: I1210 10:40:50.393777 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00d08c4e-1767-4a43-88a2-dc3e5df676f1-kube-api-access-46wjr" (OuterVolumeSpecName: "kube-api-access-46wjr") pod "00d08c4e-1767-4a43-88a2-dc3e5df676f1" (UID: "00d08c4e-1767-4a43-88a2-dc3e5df676f1"). InnerVolumeSpecName "kube-api-access-46wjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:40:50 crc kubenswrapper[4715]: I1210 10:40:50.486980 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46wjr\" (UniqueName: \"kubernetes.io/projected/00d08c4e-1767-4a43-88a2-dc3e5df676f1-kube-api-access-46wjr\") on node \"crc\" DevicePath \"\"" Dec 10 10:40:51 crc kubenswrapper[4715]: I1210 10:40:51.191494 4715 scope.go:117] "RemoveContainer" containerID="a6ab83a5729c6650b0aa88ec31d614839166d697b805833fcd37cc9dfd80bd56" Dec 10 10:40:51 crc kubenswrapper[4715]: I1210 10:40:51.191554 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/crc-debug-w72xn" Dec 10 10:40:51 crc kubenswrapper[4715]: I1210 10:40:51.617674 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00d08c4e-1767-4a43-88a2-dc3e5df676f1" path="/var/lib/kubelet/pods/00d08c4e-1767-4a43-88a2-dc3e5df676f1/volumes" Dec 10 10:41:00 crc kubenswrapper[4715]: I1210 10:41:00.604576 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:41:00 crc kubenswrapper[4715]: E1210 10:41:00.605407 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:41:10 crc kubenswrapper[4715]: I1210 10:41:10.566038 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7856564b58-9kr94_07a44832-3242-42dd-9366-5628181a6c6e/barbican-api/0.log" Dec 10 10:41:10 crc kubenswrapper[4715]: I1210 10:41:10.665667 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7856564b58-9kr94_07a44832-3242-42dd-9366-5628181a6c6e/barbican-api-log/0.log" Dec 10 10:41:10 crc kubenswrapper[4715]: I1210 10:41:10.805648 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6d588649f8-m9jx2_73357509-8bf3-408e-9c35-ee6267593be1/barbican-keystone-listener/0.log" Dec 10 10:41:10 crc kubenswrapper[4715]: I1210 10:41:10.876250 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6d588649f8-m9jx2_73357509-8bf3-408e-9c35-ee6267593be1/barbican-keystone-listener-log/0.log" Dec 10 10:41:10 crc kubenswrapper[4715]: I1210 10:41:10.996352 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c964fbd47-67rln_b46f5949-5bfa-4cfd-b70d-4d225ea12d6b/barbican-worker/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.047657 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c964fbd47-67rln_b46f5949-5bfa-4cfd-b70d-4d225ea12d6b/barbican-worker-log/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.142098 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k2frw_7bcd9508-4ca1-4bff-9e27-9c3083ee0ae2/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.320594 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3a26e430-a5f3-40cf-97fc-be1e01adb4e9/ceilometer-central-agent/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.335227 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3a26e430-a5f3-40cf-97fc-be1e01adb4e9/ceilometer-notification-agent/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.377748 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3a26e430-a5f3-40cf-97fc-be1e01adb4e9/proxy-httpd/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.460560 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3a26e430-a5f3-40cf-97fc-be1e01adb4e9/sg-core/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.566796 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e911669d-5a8e-49d6-a6e2-b66dfae8f7e5/cinder-api/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.598795 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e911669d-5a8e-49d6-a6e2-b66dfae8f7e5/cinder-api-log/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.808576 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_807957b7-0d34-49ab-8b79-6abcd56ad840/cinder-scheduler/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.838351 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_807957b7-0d34-49ab-8b79-6abcd56ad840/probe/0.log" Dec 10 10:41:11 crc kubenswrapper[4715]: I1210 10:41:11.995310 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-9r6nl_42754cce-1e7d-416d-a1fa-118ae307880f/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:12 crc kubenswrapper[4715]: I1210 10:41:12.047740 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-rxk96_c952af8b-0155-480f-a60f-399b9ec919c8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:12 crc kubenswrapper[4715]: I1210 10:41:12.215102 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-g755x_540bd567-e127-4db2-b6c5-0d447ce7b34c/init/0.log" Dec 10 10:41:12 crc kubenswrapper[4715]: I1210 10:41:12.372895 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-g755x_540bd567-e127-4db2-b6c5-0d447ce7b34c/init/0.log" Dec 10 10:41:12 crc kubenswrapper[4715]: I1210 10:41:12.401885 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-jgj4x_6d7bc0b2-e15d-4627-b901-bb6399f7f636/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:12 crc kubenswrapper[4715]: I1210 10:41:12.612961 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-g755x_540bd567-e127-4db2-b6c5-0d447ce7b34c/dnsmasq-dns/0.log" Dec 10 10:41:12 crc kubenswrapper[4715]: I1210 10:41:12.755983 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac/glance-log/0.log" Dec 10 10:41:12 crc kubenswrapper[4715]: I1210 10:41:12.774334 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_92eadfe4-ba28-475d-8e9f-4ebcb79bd9ac/glance-httpd/0.log" Dec 10 10:41:12 crc kubenswrapper[4715]: I1210 10:41:12.876166 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ba040a24-bae6-4627-a934-39d616b57c76/glance-httpd/0.log" Dec 10 10:41:12 crc kubenswrapper[4715]: I1210 10:41:12.977097 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ba040a24-bae6-4627-a934-39d616b57c76/glance-log/0.log" Dec 10 10:41:13 crc kubenswrapper[4715]: I1210 10:41:13.125706 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-857b4dfcd4-wnlbz_4214260f-423f-47fb-b062-c752f3519175/horizon/0.log" Dec 10 10:41:13 crc kubenswrapper[4715]: I1210 10:41:13.268132 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-rzlbk_0073ecf9-f0e9-4ccf-b9ab-90a9bf074bf9/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:13 crc kubenswrapper[4715]: I1210 10:41:13.551645 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-857b4dfcd4-wnlbz_4214260f-423f-47fb-b062-c752f3519175/horizon-log/0.log" Dec 10 10:41:13 crc kubenswrapper[4715]: I1210 10:41:13.605443 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:41:13 crc kubenswrapper[4715]: E1210 10:41:13.605815 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:41:14 crc kubenswrapper[4715]: I1210 10:41:14.128595 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-h82b4_9692f891-f3d4-412b-92bd-d154842e0d79/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:14 crc kubenswrapper[4715]: I1210 10:41:14.169177 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29422681-fb8s5_2992fa3e-b3d4-42da-845e-a019e8e23cf1/keystone-cron/0.log" Dec 10 10:41:14 crc kubenswrapper[4715]: I1210 10:41:14.372801 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_cf760357-b816-4b56-b56d-5a3637a7b2b0/kube-state-metrics/0.log" Dec 10 10:41:14 crc kubenswrapper[4715]: I1210 10:41:14.504984 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-d9c5c96bf-n94nb_946c4b3b-eb91-4d39-a7ef-88d25e23599e/keystone-api/0.log" Dec 10 10:41:14 crc kubenswrapper[4715]: I1210 10:41:14.531746 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-959dc_cd86f221-0c52-4565-87ed-9d4c473aaaf4/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:14 crc kubenswrapper[4715]: I1210 10:41:14.798905 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-85df69bf85-6cwpq_0cf58c4f-3594-4b20-a3fb-313c0f25a844/neutron-httpd/0.log" Dec 10 10:41:14 crc kubenswrapper[4715]: I1210 10:41:14.911613 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-zc5m6_06251a0c-f042-4137-864b-c7003449b040/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:14 crc kubenswrapper[4715]: I1210 10:41:14.928082 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-85df69bf85-6cwpq_0cf58c4f-3594-4b20-a3fb-313c0f25a844/neutron-api/0.log" Dec 10 10:41:15 crc kubenswrapper[4715]: I1210 10:41:15.478595 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2648041d-c409-4012-b90b-3d780bacb44d/nova-api-log/0.log" Dec 10 10:41:15 crc kubenswrapper[4715]: I1210 10:41:15.483198 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_6f8bdb76-8ee4-43e6-b1c6-68f4ef3814da/nova-cell0-conductor-conductor/0.log" Dec 10 10:41:15 crc kubenswrapper[4715]: I1210 10:41:15.777619 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2648041d-c409-4012-b90b-3d780bacb44d/nova-api-api/0.log" Dec 10 10:41:16 crc kubenswrapper[4715]: I1210 10:41:16.195228 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_87397cd7-eabb-4b0b-809a-c9f2ab53e8be/nova-cell1-novncproxy-novncproxy/0.log" Dec 10 10:41:16 crc kubenswrapper[4715]: I1210 10:41:16.275939 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7748551a-94b0-47c5-8360-584b386b0bf2/nova-cell1-conductor-conductor/0.log" Dec 10 10:41:16 crc kubenswrapper[4715]: I1210 10:41:16.393568 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-sbpnp_31ba5b33-f7ca-4389-879d-8a6a38cc0795/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:16 crc kubenswrapper[4715]: I1210 10:41:16.537349 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8027e8a4-9be7-4764-8c1d-f25a3eb3587b/nova-metadata-log/0.log" Dec 10 10:41:16 crc kubenswrapper[4715]: I1210 10:41:16.865524 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_6feb8a72-37ea-419b-b69e-0a2ef81d79ca/nova-scheduler-scheduler/0.log" Dec 10 10:41:16 crc kubenswrapper[4715]: I1210 10:41:16.908869 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_05dd65e3-40ea-4d12-856f-3fa4f99712b2/mysql-bootstrap/0.log" Dec 10 10:41:17 crc kubenswrapper[4715]: I1210 10:41:17.114086 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_05dd65e3-40ea-4d12-856f-3fa4f99712b2/galera/0.log" Dec 10 10:41:17 crc kubenswrapper[4715]: I1210 10:41:17.150305 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_05dd65e3-40ea-4d12-856f-3fa4f99712b2/mysql-bootstrap/0.log" Dec 10 10:41:17 crc kubenswrapper[4715]: I1210 10:41:17.341272 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8bcd31bd-4e40-41b6-92ef-6e0e0b170afa/mysql-bootstrap/0.log" Dec 10 10:41:17 crc kubenswrapper[4715]: I1210 10:41:17.543640 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8bcd31bd-4e40-41b6-92ef-6e0e0b170afa/galera/0.log" Dec 10 10:41:17 crc kubenswrapper[4715]: I1210 10:41:17.650071 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8bcd31bd-4e40-41b6-92ef-6e0e0b170afa/mysql-bootstrap/0.log" Dec 10 10:41:17 crc kubenswrapper[4715]: I1210 10:41:17.807844 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2843e1e0-1b99-45af-a422-e533ee493c78/openstackclient/0.log" Dec 10 10:41:17 crc kubenswrapper[4715]: I1210 10:41:17.892035 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-lz7jm_6fc17771-af47-4ba4-be8a-6b3143a999f4/ovn-controller/0.log" Dec 10 10:41:17 crc kubenswrapper[4715]: I1210 10:41:17.999132 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8027e8a4-9be7-4764-8c1d-f25a3eb3587b/nova-metadata-metadata/0.log" Dec 10 10:41:18 crc kubenswrapper[4715]: I1210 10:41:18.100716 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-p5m86_e26b3d5b-1e63-4d22-80a5-55ffb58f664b/openstack-network-exporter/0.log" Dec 10 10:41:18 crc kubenswrapper[4715]: I1210 10:41:18.235256 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rv2hq_d6ae191b-a288-4185-b98b-15654b1e5dbf/ovsdb-server-init/0.log" Dec 10 10:41:18 crc kubenswrapper[4715]: I1210 10:41:18.511134 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rv2hq_d6ae191b-a288-4185-b98b-15654b1e5dbf/ovsdb-server-init/0.log" Dec 10 10:41:18 crc kubenswrapper[4715]: I1210 10:41:18.511971 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rv2hq_d6ae191b-a288-4185-b98b-15654b1e5dbf/ovsdb-server/0.log" Dec 10 10:41:18 crc kubenswrapper[4715]: I1210 10:41:18.517344 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rv2hq_d6ae191b-a288-4185-b98b-15654b1e5dbf/ovs-vswitchd/0.log" Dec 10 10:41:18 crc kubenswrapper[4715]: I1210 10:41:18.714464 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b0f6351-5ab5-4e0c-ad34-95061bc14a35/openstack-network-exporter/0.log" Dec 10 10:41:18 crc kubenswrapper[4715]: I1210 10:41:18.718126 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-plxxq_b70f0fbd-ec55-40e1-b61d-b34b1b92fa3b/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:18 crc kubenswrapper[4715]: I1210 10:41:18.775904 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b0f6351-5ab5-4e0c-ad34-95061bc14a35/ovn-northd/0.log" Dec 10 10:41:18 crc kubenswrapper[4715]: I1210 10:41:18.968859 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b4bfd970-df85-4d59-a211-ba4adbb501ee/ovsdbserver-nb/0.log" Dec 10 10:41:18 crc kubenswrapper[4715]: I1210 10:41:18.973370 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b4bfd970-df85-4d59-a211-ba4adbb501ee/openstack-network-exporter/0.log" Dec 10 10:41:19 crc kubenswrapper[4715]: I1210 10:41:19.236699 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_219b8c44-8f39-40f1-9000-25404c35d495/ovsdbserver-sb/0.log" Dec 10 10:41:19 crc kubenswrapper[4715]: I1210 10:41:19.282211 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_219b8c44-8f39-40f1-9000-25404c35d495/openstack-network-exporter/0.log" Dec 10 10:41:19 crc kubenswrapper[4715]: I1210 10:41:19.345848 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5777d57b56-6lbcp_21593c1a-bc1a-4460-9ef9-3eef12261355/placement-api/0.log" Dec 10 10:41:19 crc kubenswrapper[4715]: I1210 10:41:19.528557 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf6962bf-0a77-4285-b58f-6213688e74b7/setup-container/0.log" Dec 10 10:41:19 crc kubenswrapper[4715]: I1210 10:41:19.573209 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5777d57b56-6lbcp_21593c1a-bc1a-4460-9ef9-3eef12261355/placement-log/0.log" Dec 10 10:41:19 crc kubenswrapper[4715]: I1210 10:41:19.740141 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf6962bf-0a77-4285-b58f-6213688e74b7/rabbitmq/0.log" Dec 10 10:41:19 crc kubenswrapper[4715]: I1210 10:41:19.742304 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4e8933c-d768-48cb-bac7-245abd8761ad/setup-container/0.log" Dec 10 10:41:19 crc kubenswrapper[4715]: I1210 10:41:19.777165 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf6962bf-0a77-4285-b58f-6213688e74b7/setup-container/0.log" Dec 10 10:41:20 crc kubenswrapper[4715]: I1210 10:41:20.006163 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4e8933c-d768-48cb-bac7-245abd8761ad/setup-container/0.log" Dec 10 10:41:20 crc kubenswrapper[4715]: I1210 10:41:20.039373 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4e8933c-d768-48cb-bac7-245abd8761ad/rabbitmq/0.log" Dec 10 10:41:20 crc kubenswrapper[4715]: I1210 10:41:20.074439 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-sr547_6d1a7152-aa86-4400-b3bb-55ca096ef3f7/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:20 crc kubenswrapper[4715]: I1210 10:41:20.243292 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-5rcr7_bd4fa444-6ad3-4129-b62d-c98ce1b69058/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:20 crc kubenswrapper[4715]: I1210 10:41:20.329084 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-nx5fl_17c08b14-a1de-4e79-b6d6-a8543bb0a4fd/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:20 crc kubenswrapper[4715]: I1210 10:41:20.500927 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pz4s4_06a8246c-8107-4c74-b827-1c965bd840ec/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:20 crc kubenswrapper[4715]: I1210 10:41:20.594843 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8h9tz_757f9102-c847-4fa9-b9a6-87495185cffe/ssh-known-hosts-edpm-deployment/0.log" Dec 10 10:41:20 crc kubenswrapper[4715]: I1210 10:41:20.803153 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6bcc74b4c5-ws76b_34a9b314-68fd-43d1-89c8-b55bf717f6df/proxy-server/0.log" Dec 10 10:41:20 crc kubenswrapper[4715]: I1210 10:41:20.840410 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6bcc74b4c5-ws76b_34a9b314-68fd-43d1-89c8-b55bf717f6df/proxy-httpd/0.log" Dec 10 10:41:20 crc kubenswrapper[4715]: I1210 10:41:20.903960 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-47zsh_3a351985-c466-4a2b-b74e-f67d85fb5715/swift-ring-rebalance/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.045661 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/account-auditor/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.130345 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/account-reaper/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.176041 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/account-replicator/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.235035 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/account-server/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.327857 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/container-auditor/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.365159 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/container-server/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.367012 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/container-replicator/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.491747 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/container-updater/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.574422 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/object-replicator/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.580667 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/object-auditor/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.589513 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/object-expirer/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.684272 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/object-server/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.780331 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/object-updater/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.811138 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/rsync/0.log" Dec 10 10:41:21 crc kubenswrapper[4715]: I1210 10:41:21.812794 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_de6c4b14-17ae-4701-b7bc-b4d907f6964c/swift-recon-cron/0.log" Dec 10 10:41:22 crc kubenswrapper[4715]: I1210 10:41:22.024850 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_a1dcb116-f8aa-4bbc-a28e-3a4ac2810ea8/tempest-tests-tempest-tests-runner/0.log" Dec 10 10:41:22 crc kubenswrapper[4715]: I1210 10:41:22.066776 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-v8bjk_54173a4c-eb6d-4482-9bce-3f24ab7a8b4f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:22 crc kubenswrapper[4715]: I1210 10:41:22.208767 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_40f62965-1d20-4eba-803a-b513c07e489f/test-operator-logs-container/0.log" Dec 10 10:41:22 crc kubenswrapper[4715]: I1210 10:41:22.318306 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-h2r5m_dc964814-44b2-4f7c-b4f2-7a9962758553/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 10:41:27 crc kubenswrapper[4715]: I1210 10:41:27.625782 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:41:27 crc kubenswrapper[4715]: E1210 10:41:27.626491 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:41:32 crc kubenswrapper[4715]: I1210 10:41:32.238651 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_92be40ca-d175-4ec0-819b-1579787701d7/memcached/0.log" Dec 10 10:41:42 crc kubenswrapper[4715]: I1210 10:41:42.605374 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:41:42 crc kubenswrapper[4715]: E1210 10:41:42.606150 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:41:51 crc kubenswrapper[4715]: I1210 10:41:51.829897 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/util/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.012989 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/util/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.054683 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/pull/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.066515 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/pull/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.175245 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/util/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.192429 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/pull/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.249087 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9159cec11b07029548317eae118f26998289b8aa117feb7f4de6944c11xfdqk_c077e8ab-c825-493d-b731-b3cc36eac159/extract/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.340436 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jwwm5_f1e4cba4-0694-4477-b031-5caac1fbcb84/kube-rbac-proxy/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.424775 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-jwwm5_f1e4cba4-0694-4477-b031-5caac1fbcb84/manager/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.474439 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-htklg_af548670-106c-4fbc-b6a5-102896191190/kube-rbac-proxy/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.575010 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-htklg_af548670-106c-4fbc-b6a5-102896191190/manager/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.701660 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-j2drs_5949f2c9-f510-47f4-988c-cae06576536a/kube-rbac-proxy/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.702607 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-j2drs_5949f2c9-f510-47f4-988c-cae06576536a/manager/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.883735 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-scbxh_6c14bbb7-a3cf-4654-ba2c-2d993022fa72/kube-rbac-proxy/0.log" Dec 10 10:41:52 crc kubenswrapper[4715]: I1210 10:41:52.986253 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-scbxh_6c14bbb7-a3cf-4654-ba2c-2d993022fa72/manager/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.034319 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jdcr5_c396c828-63cf-4cd7-a050-c359ce8e1c8b/kube-rbac-proxy/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.104648 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jdcr5_c396c828-63cf-4cd7-a050-c359ce8e1c8b/manager/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.184876 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-d875q_06228f2c-8f0c-493a-850f-b1e5fd41a23b/kube-rbac-proxy/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.255541 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-d875q_06228f2c-8f0c-493a-850f-b1e5fd41a23b/manager/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.407819 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-mxxcb_31619ce3-146d-4efb-898f-75bd1e007de9/kube-rbac-proxy/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.537385 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-h9rpn_c3397611-a5cb-455c-8c5e-dd24ab61545a/kube-rbac-proxy/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.606312 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-h9rpn_c3397611-a5cb-455c-8c5e-dd24ab61545a/manager/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.688658 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-mxxcb_31619ce3-146d-4efb-898f-75bd1e007de9/manager/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.788805 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-88z54_75ebabe2-8b57-4eee-9b3b-b238fafd94b7/kube-rbac-proxy/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.853725 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-88z54_75ebabe2-8b57-4eee-9b3b-b238fafd94b7/manager/0.log" Dec 10 10:41:53 crc kubenswrapper[4715]: I1210 10:41:53.970933 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-p69tc_d80992c0-9c12-4266-b705-6eaed7e46d1e/kube-rbac-proxy/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.057326 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-p69tc_d80992c0-9c12-4266-b705-6eaed7e46d1e/manager/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.135854 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-tjqbt_78ba49c9-f451-4982-8d88-6bf0444c8b60/kube-rbac-proxy/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.233720 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-tjqbt_78ba49c9-f451-4982-8d88-6bf0444c8b60/manager/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.353073 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-z25gz_20025d92-9314-436f-a05e-4bfee3d3add0/manager/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.356772 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-z25gz_20025d92-9314-436f-a05e-4bfee3d3add0/kube-rbac-proxy/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.514673 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-f8zw7_50f2991b-7b13-4d84-8a4d-504d9b873023/kube-rbac-proxy/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.646362 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-f8zw7_50f2991b-7b13-4d84-8a4d-504d9b873023/manager/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.742007 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-gbrf4_91566e3d-fbed-450a-bc66-1bf34f98d57f/manager/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.750360 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-gbrf4_91566e3d-fbed-450a-bc66-1bf34f98d57f/kube-rbac-proxy/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.838475 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fphth6_d329770a-7d39-41d0-b4b1-a52e2e9ccda7/kube-rbac-proxy/0.log" Dec 10 10:41:54 crc kubenswrapper[4715]: I1210 10:41:54.920410 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fphth6_d329770a-7d39-41d0-b4b1-a52e2e9ccda7/manager/0.log" Dec 10 10:41:55 crc kubenswrapper[4715]: I1210 10:41:55.537778 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7dcf9bf79d-7ftkw_5899529c-0209-4c8f-8b67-37c9d08bbcf8/operator/0.log" Dec 10 10:41:55 crc kubenswrapper[4715]: I1210 10:41:55.668416 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-crsjq_313f8227-63e5-435e-a2e9-0c08d67712b8/registry-server/0.log" Dec 10 10:41:55 crc kubenswrapper[4715]: I1210 10:41:55.745727 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-66fzd_01645280-6af5-4f8d-b412-3d7e0efa9054/kube-rbac-proxy/0.log" Dec 10 10:41:55 crc kubenswrapper[4715]: I1210 10:41:55.942108 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-66fzd_01645280-6af5-4f8d-b412-3d7e0efa9054/manager/0.log" Dec 10 10:41:55 crc kubenswrapper[4715]: I1210 10:41:55.988088 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-m75xm_629d31d3-4474-4beb-9c7b-2d890d2134e6/kube-rbac-proxy/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.031097 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-m75xm_629d31d3-4474-4beb-9c7b-2d890d2134e6/manager/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.206742 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-69rd4_3dccc54c-41e2-4745-839c-df51ca4e1825/operator/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.244089 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-ht6n4_67b3cab4-c74e-4861-875b-e38aac7ced71/kube-rbac-proxy/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.434045 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-ht6n4_67b3cab4-c74e-4861-875b-e38aac7ced71/manager/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.449785 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fd7979c46-c59rw_8c747bc0-5cc3-4cd2-8283-bdb2f2742167/manager/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.489767 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-2957k_63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd/kube-rbac-proxy/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.629745 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-2957k_63f8a7e2-a49d-4927-90c9-dfb3c59ffbfd/manager/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.653831 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wcvvl_cfe88ca5-7d1d-4fff-8404-8b192ae48b09/kube-rbac-proxy/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.666528 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wcvvl_cfe88ca5-7d1d-4fff-8404-8b192ae48b09/manager/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.829222 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-x98sw_d66b6939-c9a0-413b-b09a-5eb529657837/kube-rbac-proxy/0.log" Dec 10 10:41:56 crc kubenswrapper[4715]: I1210 10:41:56.890029 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-x98sw_d66b6939-c9a0-413b-b09a-5eb529657837/manager/0.log" Dec 10 10:41:57 crc kubenswrapper[4715]: I1210 10:41:57.619530 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:41:57 crc kubenswrapper[4715]: E1210 10:41:57.619860 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:42:10 crc kubenswrapper[4715]: I1210 10:42:10.605166 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:42:10 crc kubenswrapper[4715]: E1210 10:42:10.605970 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7pbhk_openshift-machine-config-operator(8b26b009-53b8-4260-88a0-0f9f9553d676)\"" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" Dec 10 10:42:17 crc kubenswrapper[4715]: I1210 10:42:17.277020 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-tnrn5_45bf2bc4-681f-46cf-aca3-93960c133428/control-plane-machine-set-operator/0.log" Dec 10 10:42:17 crc kubenswrapper[4715]: I1210 10:42:17.325052 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fd9v5_c9ce5159-4cdb-47c7-9193-8adde2d4e2d0/kube-rbac-proxy/0.log" Dec 10 10:42:17 crc kubenswrapper[4715]: I1210 10:42:17.368139 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fd9v5_c9ce5159-4cdb-47c7-9193-8adde2d4e2d0/machine-api-operator/0.log" Dec 10 10:42:21 crc kubenswrapper[4715]: I1210 10:42:21.605276 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:42:22 crc kubenswrapper[4715]: I1210 10:42:22.057355 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"af8993f362c7c351803df27f5108a49765c289c83d2af359c49536a36e4f4555"} Dec 10 10:42:31 crc kubenswrapper[4715]: I1210 10:42:31.194332 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-mmntd_7a38a68d-c3fe-40f5-be5d-4c75b919c4c4/cert-manager-controller/0.log" Dec 10 10:42:31 crc kubenswrapper[4715]: I1210 10:42:31.302509 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-cxn4s_189f8af2-11e6-4140-8104-fb4a66bce776/cert-manager-cainjector/0.log" Dec 10 10:42:31 crc kubenswrapper[4715]: I1210 10:42:31.389896 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-vckw6_7f4c1366-2e47-4899-a38f-55cbc26f0569/cert-manager-webhook/0.log" Dec 10 10:42:42 crc kubenswrapper[4715]: I1210 10:42:42.751000 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-9srkc_2993b76a-4f33-4a02-8845-e694f2f65179/nmstate-console-plugin/0.log" Dec 10 10:42:42 crc kubenswrapper[4715]: I1210 10:42:42.931666 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-hhxxc_11a0c637-5b67-4db5-8fd0-33c6fabc6103/nmstate-handler/0.log" Dec 10 10:42:42 crc kubenswrapper[4715]: I1210 10:42:42.956067 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lwx7z_adb59670-27dd-447f-8ccc-ca87956810ec/nmstate-metrics/0.log" Dec 10 10:42:42 crc kubenswrapper[4715]: I1210 10:42:42.986960 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lwx7z_adb59670-27dd-447f-8ccc-ca87956810ec/kube-rbac-proxy/0.log" Dec 10 10:42:43 crc kubenswrapper[4715]: I1210 10:42:43.210717 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-nsj47_bc0de16a-d1b8-41b0-ad21-6015ea156e9a/nmstate-webhook/0.log" Dec 10 10:42:43 crc kubenswrapper[4715]: I1210 10:42:43.226070 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-bztww_bb6d3ad9-978f-48ff-b72c-434fa2c5783e/nmstate-operator/0.log" Dec 10 10:42:58 crc kubenswrapper[4715]: I1210 10:42:58.407445 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-xsr9b_fd252c98-3f30-4b3f-aa36-4056f00035a2/kube-rbac-proxy/0.log" Dec 10 10:42:58 crc kubenswrapper[4715]: I1210 10:42:58.529043 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-xsr9b_fd252c98-3f30-4b3f-aa36-4056f00035a2/controller/0.log" Dec 10 10:42:58 crc kubenswrapper[4715]: I1210 10:42:58.668080 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-frr-files/0.log" Dec 10 10:42:58 crc kubenswrapper[4715]: I1210 10:42:58.877511 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-frr-files/0.log" Dec 10 10:42:58 crc kubenswrapper[4715]: I1210 10:42:58.907794 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-reloader/0.log" Dec 10 10:42:58 crc kubenswrapper[4715]: I1210 10:42:58.928330 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-metrics/0.log" Dec 10 10:42:58 crc kubenswrapper[4715]: I1210 10:42:58.961811 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-reloader/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.110272 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-reloader/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.132533 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-metrics/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.150362 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-frr-files/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.169726 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-metrics/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.387999 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-frr-files/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.414705 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-reloader/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.422953 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/cp-metrics/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.458090 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/controller/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.602810 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/kube-rbac-proxy/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.608369 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/frr-metrics/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.702713 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/kube-rbac-proxy-frr/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.876484 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/reloader/0.log" Dec 10 10:42:59 crc kubenswrapper[4715]: I1210 10:42:59.949654 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-s5rwg_e7f61745-e2a2-4e9a-b2a1-b931599251a8/frr-k8s-webhook-server/0.log" Dec 10 10:43:00 crc kubenswrapper[4715]: I1210 10:43:00.134597 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-69456cd765-w9pmv_6b85b47f-761c-4064-8a3e-21824cb5eef7/manager/0.log" Dec 10 10:43:00 crc kubenswrapper[4715]: I1210 10:43:00.841223 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-79dd69b788-gwkcv_711fc7a7-527a-4e3b-b343-a279ce6d3b79/webhook-server/0.log" Dec 10 10:43:00 crc kubenswrapper[4715]: I1210 10:43:00.919807 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-p2fz4_0fd6b12e-bdcd-4662-b1d4-5b2605b42295/frr/0.log" Dec 10 10:43:01 crc kubenswrapper[4715]: I1210 10:43:01.019354 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-g6jgv_6f9bf056-705e-4f97-a470-1bd98ae14e30/kube-rbac-proxy/0.log" Dec 10 10:43:01 crc kubenswrapper[4715]: I1210 10:43:01.343274 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-g6jgv_6f9bf056-705e-4f97-a470-1bd98ae14e30/speaker/0.log" Dec 10 10:43:14 crc kubenswrapper[4715]: I1210 10:43:14.365469 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/util/0.log" Dec 10 10:43:14 crc kubenswrapper[4715]: I1210 10:43:14.528755 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/pull/0.log" Dec 10 10:43:14 crc kubenswrapper[4715]: I1210 10:43:14.530707 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/util/0.log" Dec 10 10:43:14 crc kubenswrapper[4715]: I1210 10:43:14.559237 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/pull/0.log" Dec 10 10:43:14 crc kubenswrapper[4715]: I1210 10:43:14.717525 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/pull/0.log" Dec 10 10:43:14 crc kubenswrapper[4715]: I1210 10:43:14.721493 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/extract/0.log" Dec 10 10:43:14 crc kubenswrapper[4715]: I1210 10:43:14.751064 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fgqx2t_580ed0cb-7745-46e9-8437-b1e0ba6199cd/util/0.log" Dec 10 10:43:14 crc kubenswrapper[4715]: I1210 10:43:14.882632 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/util/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.096832 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/util/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.099661 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/pull/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.132191 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/pull/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.276714 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/util/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.290331 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/pull/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.335100 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8398gbj_a71d4c3d-8768-4bed-a14b-cf226090ef7d/extract/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.467362 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-utilities/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.662497 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-content/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.678105 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-utilities/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.694974 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-content/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.858602 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-utilities/0.log" Dec 10 10:43:15 crc kubenswrapper[4715]: I1210 10:43:15.864699 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/extract-content/0.log" Dec 10 10:43:16 crc kubenswrapper[4715]: I1210 10:43:16.019945 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6lhbn_bc172f3d-d7ab-4675-8720-107c38a9f90a/registry-server/0.log" Dec 10 10:43:16 crc kubenswrapper[4715]: I1210 10:43:16.102447 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-utilities/0.log" Dec 10 10:43:16 crc kubenswrapper[4715]: I1210 10:43:16.256199 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-utilities/0.log" Dec 10 10:43:16 crc kubenswrapper[4715]: I1210 10:43:16.306517 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-content/0.log" Dec 10 10:43:16 crc kubenswrapper[4715]: I1210 10:43:16.310080 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-content/0.log" Dec 10 10:43:16 crc kubenswrapper[4715]: I1210 10:43:16.426394 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-utilities/0.log" Dec 10 10:43:16 crc kubenswrapper[4715]: I1210 10:43:16.436782 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/extract-content/0.log" Dec 10 10:43:16 crc kubenswrapper[4715]: I1210 10:43:16.734049 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-rjxvk_ed4c45ce-d179-4515-8166-9f1c6c7e0913/marketplace-operator/0.log" Dec 10 10:43:16 crc kubenswrapper[4715]: I1210 10:43:16.848796 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-utilities/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.015088 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-content/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.038396 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-utilities/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.159965 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nv8h7_956621e9-dfe5-4b87-97c8-458b6f3c5b6d/registry-server/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.171707 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-content/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.315023 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-content/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.316714 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/extract-utilities/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.439826 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ffpj_2ea9a921-7745-4485-8ba2-b94302d357ad/registry-server/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.520306 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-utilities/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.669310 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-utilities/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.689269 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-content/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.715231 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-content/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.856177 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-utilities/0.log" Dec 10 10:43:17 crc kubenswrapper[4715]: I1210 10:43:17.887215 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/extract-content/0.log" Dec 10 10:43:18 crc kubenswrapper[4715]: I1210 10:43:18.344825 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hb8cn_63a49347-3b45-4bd9-a158-6deb79a28d2d/registry-server/0.log" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.152406 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9wszk"] Dec 10 10:44:37 crc kubenswrapper[4715]: E1210 10:44:37.153983 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d08c4e-1767-4a43-88a2-dc3e5df676f1" containerName="container-00" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.154004 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d08c4e-1767-4a43-88a2-dc3e5df676f1" containerName="container-00" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.154272 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d08c4e-1767-4a43-88a2-dc3e5df676f1" containerName="container-00" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.156213 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.178629 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9wszk"] Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.204380 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-utilities\") pod \"redhat-marketplace-9wszk\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.204467 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg6p2\" (UniqueName: \"kubernetes.io/projected/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-kube-api-access-zg6p2\") pod \"redhat-marketplace-9wszk\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.204498 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-catalog-content\") pod \"redhat-marketplace-9wszk\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.306880 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-utilities\") pod \"redhat-marketplace-9wszk\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.306993 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg6p2\" (UniqueName: \"kubernetes.io/projected/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-kube-api-access-zg6p2\") pod \"redhat-marketplace-9wszk\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.307028 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-catalog-content\") pod \"redhat-marketplace-9wszk\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.307402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-utilities\") pod \"redhat-marketplace-9wszk\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.307463 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-catalog-content\") pod \"redhat-marketplace-9wszk\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.335834 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg6p2\" (UniqueName: \"kubernetes.io/projected/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-kube-api-access-zg6p2\") pod \"redhat-marketplace-9wszk\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:37 crc kubenswrapper[4715]: I1210 10:44:37.506000 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:38 crc kubenswrapper[4715]: I1210 10:44:38.031663 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9wszk"] Dec 10 10:44:38 crc kubenswrapper[4715]: I1210 10:44:38.368267 4715 generic.go:334] "Generic (PLEG): container finished" podID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerID="a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5" exitCode=0 Dec 10 10:44:38 crc kubenswrapper[4715]: I1210 10:44:38.368319 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wszk" event={"ID":"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b","Type":"ContainerDied","Data":"a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5"} Dec 10 10:44:38 crc kubenswrapper[4715]: I1210 10:44:38.368371 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wszk" event={"ID":"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b","Type":"ContainerStarted","Data":"831c885c7fb92b5ca290664d1b0e2e0e0a9d90059df7d354d13c943b038d3711"} Dec 10 10:44:38 crc kubenswrapper[4715]: I1210 10:44:38.373880 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.379533 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wszk" event={"ID":"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b","Type":"ContainerStarted","Data":"98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098"} Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.734230 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rpfjz"] Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.736269 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.749679 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rpfjz"] Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.771364 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-catalog-content\") pod \"community-operators-rpfjz\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.771437 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbng4\" (UniqueName: \"kubernetes.io/projected/481b5dc9-6323-4938-a300-45b98b999dd7-kube-api-access-fbng4\") pod \"community-operators-rpfjz\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.771556 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-utilities\") pod \"community-operators-rpfjz\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.873834 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-utilities\") pod \"community-operators-rpfjz\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.874107 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-catalog-content\") pod \"community-operators-rpfjz\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.874180 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbng4\" (UniqueName: \"kubernetes.io/projected/481b5dc9-6323-4938-a300-45b98b999dd7-kube-api-access-fbng4\") pod \"community-operators-rpfjz\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.874666 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-utilities\") pod \"community-operators-rpfjz\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:39 crc kubenswrapper[4715]: I1210 10:44:39.874793 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-catalog-content\") pod \"community-operators-rpfjz\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:40 crc kubenswrapper[4715]: I1210 10:44:40.244298 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbng4\" (UniqueName: \"kubernetes.io/projected/481b5dc9-6323-4938-a300-45b98b999dd7-kube-api-access-fbng4\") pod \"community-operators-rpfjz\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:40 crc kubenswrapper[4715]: I1210 10:44:40.373407 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:44:40 crc kubenswrapper[4715]: I1210 10:44:40.409086 4715 generic.go:334] "Generic (PLEG): container finished" podID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerID="98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098" exitCode=0 Dec 10 10:44:40 crc kubenswrapper[4715]: I1210 10:44:40.409219 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wszk" event={"ID":"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b","Type":"ContainerDied","Data":"98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098"} Dec 10 10:44:40 crc kubenswrapper[4715]: I1210 10:44:40.892683 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rpfjz"] Dec 10 10:44:40 crc kubenswrapper[4715]: W1210 10:44:40.899602 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod481b5dc9_6323_4938_a300_45b98b999dd7.slice/crio-8c900cd9ec4ff9db5828406d1b50469965be8ae815b1d4598a8e6afe745db06a WatchSource:0}: Error finding container 8c900cd9ec4ff9db5828406d1b50469965be8ae815b1d4598a8e6afe745db06a: Status 404 returned error can't find the container with id 8c900cd9ec4ff9db5828406d1b50469965be8ae815b1d4598a8e6afe745db06a Dec 10 10:44:41 crc kubenswrapper[4715]: I1210 10:44:41.418030 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpfjz" event={"ID":"481b5dc9-6323-4938-a300-45b98b999dd7","Type":"ContainerStarted","Data":"8c900cd9ec4ff9db5828406d1b50469965be8ae815b1d4598a8e6afe745db06a"} Dec 10 10:44:42 crc kubenswrapper[4715]: I1210 10:44:42.428105 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpfjz" event={"ID":"481b5dc9-6323-4938-a300-45b98b999dd7","Type":"ContainerStarted","Data":"33013bf1f7a791dd8695cc180aa28b47bf2164d38c91d2bdf266dc1b70fd5f8a"} Dec 10 10:44:42 crc kubenswrapper[4715]: I1210 10:44:42.431250 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wszk" event={"ID":"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b","Type":"ContainerStarted","Data":"e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab"} Dec 10 10:44:42 crc kubenswrapper[4715]: I1210 10:44:42.474854 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9wszk" podStartSLOduration=3.026353255 podStartE2EDuration="5.474835317s" podCreationTimestamp="2025-12-10 10:44:37 +0000 UTC" firstStartedPulling="2025-12-10 10:44:38.373552999 +0000 UTC m=+4241.117099250" lastFinishedPulling="2025-12-10 10:44:40.822035061 +0000 UTC m=+4243.565581312" observedRunningTime="2025-12-10 10:44:42.471822631 +0000 UTC m=+4245.215368882" watchObservedRunningTime="2025-12-10 10:44:42.474835317 +0000 UTC m=+4245.218381568" Dec 10 10:44:43 crc kubenswrapper[4715]: I1210 10:44:43.445441 4715 generic.go:334] "Generic (PLEG): container finished" podID="481b5dc9-6323-4938-a300-45b98b999dd7" containerID="33013bf1f7a791dd8695cc180aa28b47bf2164d38c91d2bdf266dc1b70fd5f8a" exitCode=0 Dec 10 10:44:43 crc kubenswrapper[4715]: I1210 10:44:43.445504 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpfjz" event={"ID":"481b5dc9-6323-4938-a300-45b98b999dd7","Type":"ContainerDied","Data":"33013bf1f7a791dd8695cc180aa28b47bf2164d38c91d2bdf266dc1b70fd5f8a"} Dec 10 10:44:47 crc kubenswrapper[4715]: I1210 10:44:47.506541 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:47 crc kubenswrapper[4715]: I1210 10:44:47.507162 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:47 crc kubenswrapper[4715]: I1210 10:44:47.572680 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:47 crc kubenswrapper[4715]: I1210 10:44:47.714071 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:44:47 crc kubenswrapper[4715]: I1210 10:44:47.714154 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:44:48 crc kubenswrapper[4715]: I1210 10:44:48.502624 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpfjz" event={"ID":"481b5dc9-6323-4938-a300-45b98b999dd7","Type":"ContainerStarted","Data":"485da2d12641577337bc28e22444431d54d4660598ecee52c5b686d94ac78499"} Dec 10 10:44:48 crc kubenswrapper[4715]: I1210 10:44:48.577633 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:49 crc kubenswrapper[4715]: I1210 10:44:49.512803 4715 generic.go:334] "Generic (PLEG): container finished" podID="481b5dc9-6323-4938-a300-45b98b999dd7" containerID="485da2d12641577337bc28e22444431d54d4660598ecee52c5b686d94ac78499" exitCode=0 Dec 10 10:44:49 crc kubenswrapper[4715]: I1210 10:44:49.512905 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpfjz" event={"ID":"481b5dc9-6323-4938-a300-45b98b999dd7","Type":"ContainerDied","Data":"485da2d12641577337bc28e22444431d54d4660598ecee52c5b686d94ac78499"} Dec 10 10:44:50 crc kubenswrapper[4715]: I1210 10:44:50.533497 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpfjz" event={"ID":"481b5dc9-6323-4938-a300-45b98b999dd7","Type":"ContainerStarted","Data":"0a4467b1c2b47a40aa55b1b407304e821f480ea61cb1dd0ecfe51ddabe95d11e"} Dec 10 10:44:50 crc kubenswrapper[4715]: I1210 10:44:50.536220 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9wszk"] Dec 10 10:44:50 crc kubenswrapper[4715]: I1210 10:44:50.536498 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9wszk" podUID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerName="registry-server" containerID="cri-o://e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab" gracePeriod=2 Dec 10 10:44:50 crc kubenswrapper[4715]: I1210 10:44:50.556322 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rpfjz" podStartSLOduration=4.91969612 podStartE2EDuration="11.556301453s" podCreationTimestamp="2025-12-10 10:44:39 +0000 UTC" firstStartedPulling="2025-12-10 10:44:43.447662176 +0000 UTC m=+4246.191208427" lastFinishedPulling="2025-12-10 10:44:50.084267509 +0000 UTC m=+4252.827813760" observedRunningTime="2025-12-10 10:44:50.552101433 +0000 UTC m=+4253.295647684" watchObservedRunningTime="2025-12-10 10:44:50.556301453 +0000 UTC m=+4253.299847714" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.047570 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.125314 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg6p2\" (UniqueName: \"kubernetes.io/projected/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-kube-api-access-zg6p2\") pod \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.125409 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-catalog-content\") pod \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.125539 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-utilities\") pod \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\" (UID: \"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b\") " Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.126733 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-utilities" (OuterVolumeSpecName: "utilities") pod "3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" (UID: "3cf303f6-87f2-499f-a3bb-c2f01e32ea4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.132314 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-kube-api-access-zg6p2" (OuterVolumeSpecName: "kube-api-access-zg6p2") pod "3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" (UID: "3cf303f6-87f2-499f-a3bb-c2f01e32ea4b"). InnerVolumeSpecName "kube-api-access-zg6p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.154447 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" (UID: "3cf303f6-87f2-499f-a3bb-c2f01e32ea4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.228387 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg6p2\" (UniqueName: \"kubernetes.io/projected/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-kube-api-access-zg6p2\") on node \"crc\" DevicePath \"\"" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.228436 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.228447 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.543936 4715 generic.go:334] "Generic (PLEG): container finished" podID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerID="e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab" exitCode=0 Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.543991 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9wszk" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.544030 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wszk" event={"ID":"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b","Type":"ContainerDied","Data":"e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab"} Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.544105 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wszk" event={"ID":"3cf303f6-87f2-499f-a3bb-c2f01e32ea4b","Type":"ContainerDied","Data":"831c885c7fb92b5ca290664d1b0e2e0e0a9d90059df7d354d13c943b038d3711"} Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.544131 4715 scope.go:117] "RemoveContainer" containerID="e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.588355 4715 scope.go:117] "RemoveContainer" containerID="98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098" Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.590112 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9wszk"] Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.601467 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9wszk"] Dec 10 10:44:51 crc kubenswrapper[4715]: I1210 10:44:51.617665 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" path="/var/lib/kubelet/pods/3cf303f6-87f2-499f-a3bb-c2f01e32ea4b/volumes" Dec 10 10:44:52 crc kubenswrapper[4715]: I1210 10:44:52.054363 4715 scope.go:117] "RemoveContainer" containerID="a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5" Dec 10 10:44:52 crc kubenswrapper[4715]: I1210 10:44:52.106174 4715 scope.go:117] "RemoveContainer" containerID="e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab" Dec 10 10:44:52 crc kubenswrapper[4715]: E1210 10:44:52.107447 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab\": container with ID starting with e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab not found: ID does not exist" containerID="e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab" Dec 10 10:44:52 crc kubenswrapper[4715]: I1210 10:44:52.107497 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab"} err="failed to get container status \"e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab\": rpc error: code = NotFound desc = could not find container \"e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab\": container with ID starting with e5aa769e4047838162dd818503b9aca2133b0b534b011b7a56808c909b17e4ab not found: ID does not exist" Dec 10 10:44:52 crc kubenswrapper[4715]: I1210 10:44:52.107528 4715 scope.go:117] "RemoveContainer" containerID="98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098" Dec 10 10:44:52 crc kubenswrapper[4715]: E1210 10:44:52.108024 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098\": container with ID starting with 98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098 not found: ID does not exist" containerID="98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098" Dec 10 10:44:52 crc kubenswrapper[4715]: I1210 10:44:52.108094 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098"} err="failed to get container status \"98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098\": rpc error: code = NotFound desc = could not find container \"98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098\": container with ID starting with 98a1772041a9793f5be6327ac6f63983ca9833164c0d44363275fda593b99098 not found: ID does not exist" Dec 10 10:44:52 crc kubenswrapper[4715]: I1210 10:44:52.108136 4715 scope.go:117] "RemoveContainer" containerID="a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5" Dec 10 10:44:52 crc kubenswrapper[4715]: E1210 10:44:52.108644 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5\": container with ID starting with a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5 not found: ID does not exist" containerID="a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5" Dec 10 10:44:52 crc kubenswrapper[4715]: I1210 10:44:52.108681 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5"} err="failed to get container status \"a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5\": rpc error: code = NotFound desc = could not find container \"a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5\": container with ID starting with a5566d5485d217513bd7968a35ed2bd17d9be3816cef6326b244d63b82912aa5 not found: ID does not exist" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.193565 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5"] Dec 10 10:45:00 crc kubenswrapper[4715]: E1210 10:45:00.194469 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerName="extract-utilities" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.194482 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerName="extract-utilities" Dec 10 10:45:00 crc kubenswrapper[4715]: E1210 10:45:00.194494 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerName="registry-server" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.194501 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerName="registry-server" Dec 10 10:45:00 crc kubenswrapper[4715]: E1210 10:45:00.194523 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerName="extract-content" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.194529 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerName="extract-content" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.194857 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf303f6-87f2-499f-a3bb-c2f01e32ea4b" containerName="registry-server" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.195567 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.198381 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.204313 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5"] Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.205756 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.340182 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d6b1960-9fec-49da-878c-afcf2a435ad2-secret-volume\") pod \"collect-profiles-29422725-5nmq5\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.340236 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d6b1960-9fec-49da-878c-afcf2a435ad2-config-volume\") pod \"collect-profiles-29422725-5nmq5\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.340280 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmvsv\" (UniqueName: \"kubernetes.io/projected/1d6b1960-9fec-49da-878c-afcf2a435ad2-kube-api-access-kmvsv\") pod \"collect-profiles-29422725-5nmq5\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.374835 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.374964 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.422945 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.443202 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmvsv\" (UniqueName: \"kubernetes.io/projected/1d6b1960-9fec-49da-878c-afcf2a435ad2-kube-api-access-kmvsv\") pod \"collect-profiles-29422725-5nmq5\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.443413 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d6b1960-9fec-49da-878c-afcf2a435ad2-secret-volume\") pod \"collect-profiles-29422725-5nmq5\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.443446 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d6b1960-9fec-49da-878c-afcf2a435ad2-config-volume\") pod \"collect-profiles-29422725-5nmq5\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.444374 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d6b1960-9fec-49da-878c-afcf2a435ad2-config-volume\") pod \"collect-profiles-29422725-5nmq5\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.459329 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d6b1960-9fec-49da-878c-afcf2a435ad2-secret-volume\") pod \"collect-profiles-29422725-5nmq5\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.460900 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmvsv\" (UniqueName: \"kubernetes.io/projected/1d6b1960-9fec-49da-878c-afcf2a435ad2-kube-api-access-kmvsv\") pod \"collect-profiles-29422725-5nmq5\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.530447 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.685827 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.751235 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rpfjz"] Dec 10 10:45:00 crc kubenswrapper[4715]: W1210 10:45:00.990479 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d6b1960_9fec_49da_878c_afcf2a435ad2.slice/crio-51c15ce45d6a133d1cd74bc575ece5ff964d583a18bf5cc3116a914b60caa6bf WatchSource:0}: Error finding container 51c15ce45d6a133d1cd74bc575ece5ff964d583a18bf5cc3116a914b60caa6bf: Status 404 returned error can't find the container with id 51c15ce45d6a133d1cd74bc575ece5ff964d583a18bf5cc3116a914b60caa6bf Dec 10 10:45:00 crc kubenswrapper[4715]: I1210 10:45:00.995240 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5"] Dec 10 10:45:01 crc kubenswrapper[4715]: I1210 10:45:01.628836 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" event={"ID":"1d6b1960-9fec-49da-878c-afcf2a435ad2","Type":"ContainerStarted","Data":"7062eced531e525137af4f59201932c6ada17f96d6a6e2ef6910d2bf4e63cbff"} Dec 10 10:45:01 crc kubenswrapper[4715]: I1210 10:45:01.629589 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" event={"ID":"1d6b1960-9fec-49da-878c-afcf2a435ad2","Type":"ContainerStarted","Data":"51c15ce45d6a133d1cd74bc575ece5ff964d583a18bf5cc3116a914b60caa6bf"} Dec 10 10:45:01 crc kubenswrapper[4715]: I1210 10:45:01.651369 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" podStartSLOduration=1.6513440240000001 podStartE2EDuration="1.651344024s" podCreationTimestamp="2025-12-10 10:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 10:45:01.644655503 +0000 UTC m=+4264.388201754" watchObservedRunningTime="2025-12-10 10:45:01.651344024 +0000 UTC m=+4264.394890275" Dec 10 10:45:02 crc kubenswrapper[4715]: I1210 10:45:02.641332 4715 generic.go:334] "Generic (PLEG): container finished" podID="1d6b1960-9fec-49da-878c-afcf2a435ad2" containerID="7062eced531e525137af4f59201932c6ada17f96d6a6e2ef6910d2bf4e63cbff" exitCode=0 Dec 10 10:45:02 crc kubenswrapper[4715]: I1210 10:45:02.641759 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" event={"ID":"1d6b1960-9fec-49da-878c-afcf2a435ad2","Type":"ContainerDied","Data":"7062eced531e525137af4f59201932c6ada17f96d6a6e2ef6910d2bf4e63cbff"} Dec 10 10:45:02 crc kubenswrapper[4715]: I1210 10:45:02.647794 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rpfjz" podUID="481b5dc9-6323-4938-a300-45b98b999dd7" containerName="registry-server" containerID="cri-o://0a4467b1c2b47a40aa55b1b407304e821f480ea61cb1dd0ecfe51ddabe95d11e" gracePeriod=2 Dec 10 10:45:03 crc kubenswrapper[4715]: I1210 10:45:03.663244 4715 generic.go:334] "Generic (PLEG): container finished" podID="481b5dc9-6323-4938-a300-45b98b999dd7" containerID="0a4467b1c2b47a40aa55b1b407304e821f480ea61cb1dd0ecfe51ddabe95d11e" exitCode=0 Dec 10 10:45:03 crc kubenswrapper[4715]: I1210 10:45:03.663434 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpfjz" event={"ID":"481b5dc9-6323-4938-a300-45b98b999dd7","Type":"ContainerDied","Data":"0a4467b1c2b47a40aa55b1b407304e821f480ea61cb1dd0ecfe51ddabe95d11e"} Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.036611 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.040984 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.136967 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-catalog-content\") pod \"481b5dc9-6323-4938-a300-45b98b999dd7\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.137122 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d6b1960-9fec-49da-878c-afcf2a435ad2-secret-volume\") pod \"1d6b1960-9fec-49da-878c-afcf2a435ad2\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.137285 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmvsv\" (UniqueName: \"kubernetes.io/projected/1d6b1960-9fec-49da-878c-afcf2a435ad2-kube-api-access-kmvsv\") pod \"1d6b1960-9fec-49da-878c-afcf2a435ad2\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.137369 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbng4\" (UniqueName: \"kubernetes.io/projected/481b5dc9-6323-4938-a300-45b98b999dd7-kube-api-access-fbng4\") pod \"481b5dc9-6323-4938-a300-45b98b999dd7\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.137395 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-utilities\") pod \"481b5dc9-6323-4938-a300-45b98b999dd7\" (UID: \"481b5dc9-6323-4938-a300-45b98b999dd7\") " Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.137418 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d6b1960-9fec-49da-878c-afcf2a435ad2-config-volume\") pod \"1d6b1960-9fec-49da-878c-afcf2a435ad2\" (UID: \"1d6b1960-9fec-49da-878c-afcf2a435ad2\") " Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.139170 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d6b1960-9fec-49da-878c-afcf2a435ad2-config-volume" (OuterVolumeSpecName: "config-volume") pod "1d6b1960-9fec-49da-878c-afcf2a435ad2" (UID: "1d6b1960-9fec-49da-878c-afcf2a435ad2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.139513 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-utilities" (OuterVolumeSpecName: "utilities") pod "481b5dc9-6323-4938-a300-45b98b999dd7" (UID: "481b5dc9-6323-4938-a300-45b98b999dd7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.146125 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/481b5dc9-6323-4938-a300-45b98b999dd7-kube-api-access-fbng4" (OuterVolumeSpecName: "kube-api-access-fbng4") pod "481b5dc9-6323-4938-a300-45b98b999dd7" (UID: "481b5dc9-6323-4938-a300-45b98b999dd7"). InnerVolumeSpecName "kube-api-access-fbng4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.146601 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d6b1960-9fec-49da-878c-afcf2a435ad2-kube-api-access-kmvsv" (OuterVolumeSpecName: "kube-api-access-kmvsv") pod "1d6b1960-9fec-49da-878c-afcf2a435ad2" (UID: "1d6b1960-9fec-49da-878c-afcf2a435ad2"). InnerVolumeSpecName "kube-api-access-kmvsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.158508 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d6b1960-9fec-49da-878c-afcf2a435ad2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1d6b1960-9fec-49da-878c-afcf2a435ad2" (UID: "1d6b1960-9fec-49da-878c-afcf2a435ad2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.188633 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "481b5dc9-6323-4938-a300-45b98b999dd7" (UID: "481b5dc9-6323-4938-a300-45b98b999dd7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.239804 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.239847 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d6b1960-9fec-49da-878c-afcf2a435ad2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.239877 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmvsv\" (UniqueName: \"kubernetes.io/projected/1d6b1960-9fec-49da-878c-afcf2a435ad2-kube-api-access-kmvsv\") on node \"crc\" DevicePath \"\"" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.239890 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbng4\" (UniqueName: \"kubernetes.io/projected/481b5dc9-6323-4938-a300-45b98b999dd7-kube-api-access-fbng4\") on node \"crc\" DevicePath \"\"" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.239901 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/481b5dc9-6323-4938-a300-45b98b999dd7-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.239932 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d6b1960-9fec-49da-878c-afcf2a435ad2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.676793 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" event={"ID":"1d6b1960-9fec-49da-878c-afcf2a435ad2","Type":"ContainerDied","Data":"51c15ce45d6a133d1cd74bc575ece5ff964d583a18bf5cc3116a914b60caa6bf"} Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.677154 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51c15ce45d6a133d1cd74bc575ece5ff964d583a18bf5cc3116a914b60caa6bf" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.677229 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422725-5nmq5" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.693890 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpfjz" event={"ID":"481b5dc9-6323-4938-a300-45b98b999dd7","Type":"ContainerDied","Data":"8c900cd9ec4ff9db5828406d1b50469965be8ae815b1d4598a8e6afe745db06a"} Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.693980 4715 scope.go:117] "RemoveContainer" containerID="0a4467b1c2b47a40aa55b1b407304e821f480ea61cb1dd0ecfe51ddabe95d11e" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.693989 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpfjz" Dec 10 10:45:04 crc kubenswrapper[4715]: I1210 10:45:04.723799 4715 scope.go:117] "RemoveContainer" containerID="485da2d12641577337bc28e22444431d54d4660598ecee52c5b686d94ac78499" Dec 10 10:45:05 crc kubenswrapper[4715]: I1210 10:45:04.727025 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4"] Dec 10 10:45:05 crc kubenswrapper[4715]: I1210 10:45:04.742010 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422680-6hjk4"] Dec 10 10:45:05 crc kubenswrapper[4715]: I1210 10:45:04.751572 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rpfjz"] Dec 10 10:45:05 crc kubenswrapper[4715]: I1210 10:45:04.760161 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rpfjz"] Dec 10 10:45:05 crc kubenswrapper[4715]: I1210 10:45:04.768293 4715 scope.go:117] "RemoveContainer" containerID="33013bf1f7a791dd8695cc180aa28b47bf2164d38c91d2bdf266dc1b70fd5f8a" Dec 10 10:45:05 crc kubenswrapper[4715]: I1210 10:45:05.619484 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="481b5dc9-6323-4938-a300-45b98b999dd7" path="/var/lib/kubelet/pods/481b5dc9-6323-4938-a300-45b98b999dd7/volumes" Dec 10 10:45:05 crc kubenswrapper[4715]: I1210 10:45:05.620806 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb60d0d6-63d3-4453-9ae4-365ff52a48de" path="/var/lib/kubelet/pods/eb60d0d6-63d3-4453-9ae4-365ff52a48de/volumes" Dec 10 10:45:05 crc kubenswrapper[4715]: I1210 10:45:05.710024 4715 generic.go:334] "Generic (PLEG): container finished" podID="8e73665f-3e61-4eda-8069-e05d39a614fc" containerID="0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e" exitCode=0 Dec 10 10:45:05 crc kubenswrapper[4715]: I1210 10:45:05.710126 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nr8ss/must-gather-4spj6" event={"ID":"8e73665f-3e61-4eda-8069-e05d39a614fc","Type":"ContainerDied","Data":"0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e"} Dec 10 10:45:05 crc kubenswrapper[4715]: I1210 10:45:05.710782 4715 scope.go:117] "RemoveContainer" containerID="0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e" Dec 10 10:45:06 crc kubenswrapper[4715]: I1210 10:45:06.236891 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nr8ss_must-gather-4spj6_8e73665f-3e61-4eda-8069-e05d39a614fc/gather/0.log" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.004215 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nr8ss/must-gather-4spj6"] Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.005203 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-nr8ss/must-gather-4spj6" podUID="8e73665f-3e61-4eda-8069-e05d39a614fc" containerName="copy" containerID="cri-o://f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25" gracePeriod=2 Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.026852 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nr8ss/must-gather-4spj6"] Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.475460 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nr8ss_must-gather-4spj6_8e73665f-3e61-4eda-8069-e05d39a614fc/copy/0.log" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.476104 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/must-gather-4spj6" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.614567 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8e73665f-3e61-4eda-8069-e05d39a614fc-must-gather-output\") pod \"8e73665f-3e61-4eda-8069-e05d39a614fc\" (UID: \"8e73665f-3e61-4eda-8069-e05d39a614fc\") " Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.614728 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94pjt\" (UniqueName: \"kubernetes.io/projected/8e73665f-3e61-4eda-8069-e05d39a614fc-kube-api-access-94pjt\") pod \"8e73665f-3e61-4eda-8069-e05d39a614fc\" (UID: \"8e73665f-3e61-4eda-8069-e05d39a614fc\") " Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.621371 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e73665f-3e61-4eda-8069-e05d39a614fc-kube-api-access-94pjt" (OuterVolumeSpecName: "kube-api-access-94pjt") pod "8e73665f-3e61-4eda-8069-e05d39a614fc" (UID: "8e73665f-3e61-4eda-8069-e05d39a614fc"). InnerVolumeSpecName "kube-api-access-94pjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.714057 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.714107 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.717596 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94pjt\" (UniqueName: \"kubernetes.io/projected/8e73665f-3e61-4eda-8069-e05d39a614fc-kube-api-access-94pjt\") on node \"crc\" DevicePath \"\"" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.764469 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e73665f-3e61-4eda-8069-e05d39a614fc-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "8e73665f-3e61-4eda-8069-e05d39a614fc" (UID: "8e73665f-3e61-4eda-8069-e05d39a614fc"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.820256 4715 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8e73665f-3e61-4eda-8069-e05d39a614fc-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.828015 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nr8ss_must-gather-4spj6_8e73665f-3e61-4eda-8069-e05d39a614fc/copy/0.log" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.828475 4715 generic.go:334] "Generic (PLEG): container finished" podID="8e73665f-3e61-4eda-8069-e05d39a614fc" containerID="f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25" exitCode=143 Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.828569 4715 scope.go:117] "RemoveContainer" containerID="f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.828884 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nr8ss/must-gather-4spj6" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.848826 4715 scope.go:117] "RemoveContainer" containerID="0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.930377 4715 scope.go:117] "RemoveContainer" containerID="f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25" Dec 10 10:45:17 crc kubenswrapper[4715]: E1210 10:45:17.930895 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25\": container with ID starting with f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25 not found: ID does not exist" containerID="f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.930976 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25"} err="failed to get container status \"f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25\": rpc error: code = NotFound desc = could not find container \"f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25\": container with ID starting with f059c1885b8acabf101a51921b564c264b2ee09bafb206c0277b35c816794a25 not found: ID does not exist" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.931007 4715 scope.go:117] "RemoveContainer" containerID="0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e" Dec 10 10:45:17 crc kubenswrapper[4715]: E1210 10:45:17.931429 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e\": container with ID starting with 0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e not found: ID does not exist" containerID="0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e" Dec 10 10:45:17 crc kubenswrapper[4715]: I1210 10:45:17.931457 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e"} err="failed to get container status \"0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e\": rpc error: code = NotFound desc = could not find container \"0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e\": container with ID starting with 0c53c0fc2566c38a5bf52235965abae7deb75f6aa99bd2feaff7d996b176b84e not found: ID does not exist" Dec 10 10:45:19 crc kubenswrapper[4715]: I1210 10:45:19.614582 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e73665f-3e61-4eda-8069-e05d39a614fc" path="/var/lib/kubelet/pods/8e73665f-3e61-4eda-8069-e05d39a614fc/volumes" Dec 10 10:45:33 crc kubenswrapper[4715]: I1210 10:45:33.415816 4715 scope.go:117] "RemoveContainer" containerID="fcdb2b908a243bc0856c2c69201d5ff42d91c4c923a59777d9c1326f8f4a18d9" Dec 10 10:45:47 crc kubenswrapper[4715]: I1210 10:45:47.714110 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:45:47 crc kubenswrapper[4715]: I1210 10:45:47.714786 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 10:45:47 crc kubenswrapper[4715]: I1210 10:45:47.714842 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" Dec 10 10:45:47 crc kubenswrapper[4715]: I1210 10:45:47.716483 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af8993f362c7c351803df27f5108a49765c289c83d2af359c49536a36e4f4555"} pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 10:45:47 crc kubenswrapper[4715]: I1210 10:45:47.716677 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" containerID="cri-o://af8993f362c7c351803df27f5108a49765c289c83d2af359c49536a36e4f4555" gracePeriod=600 Dec 10 10:45:48 crc kubenswrapper[4715]: I1210 10:45:48.094824 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerID="af8993f362c7c351803df27f5108a49765c289c83d2af359c49536a36e4f4555" exitCode=0 Dec 10 10:45:48 crc kubenswrapper[4715]: I1210 10:45:48.095034 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerDied","Data":"af8993f362c7c351803df27f5108a49765c289c83d2af359c49536a36e4f4555"} Dec 10 10:45:48 crc kubenswrapper[4715]: I1210 10:45:48.095221 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" event={"ID":"8b26b009-53b8-4260-88a0-0f9f9553d676","Type":"ContainerStarted","Data":"4f415ba429ef134813f57356c3eb4d3899eee8105a18889e1622d290ad487604"} Dec 10 10:45:48 crc kubenswrapper[4715]: I1210 10:45:48.095252 4715 scope.go:117] "RemoveContainer" containerID="6dad18534c1321ec74f380a04e9c97eb0f0bee88d75e7d4a1fffbd9ea7b959e5" Dec 10 10:46:33 crc kubenswrapper[4715]: I1210 10:46:33.555938 4715 scope.go:117] "RemoveContainer" containerID="85126296ad76da71449282306d240cbb394ef0846f0478ef4302df17505c59da" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.495484 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7w2x7"] Dec 10 10:47:11 crc kubenswrapper[4715]: E1210 10:47:11.504477 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="481b5dc9-6323-4938-a300-45b98b999dd7" containerName="extract-content" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.504499 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="481b5dc9-6323-4938-a300-45b98b999dd7" containerName="extract-content" Dec 10 10:47:11 crc kubenswrapper[4715]: E1210 10:47:11.504522 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e73665f-3e61-4eda-8069-e05d39a614fc" containerName="gather" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.504530 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e73665f-3e61-4eda-8069-e05d39a614fc" containerName="gather" Dec 10 10:47:11 crc kubenswrapper[4715]: E1210 10:47:11.504551 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="481b5dc9-6323-4938-a300-45b98b999dd7" containerName="extract-utilities" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.504559 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="481b5dc9-6323-4938-a300-45b98b999dd7" containerName="extract-utilities" Dec 10 10:47:11 crc kubenswrapper[4715]: E1210 10:47:11.504572 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d6b1960-9fec-49da-878c-afcf2a435ad2" containerName="collect-profiles" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.504580 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d6b1960-9fec-49da-878c-afcf2a435ad2" containerName="collect-profiles" Dec 10 10:47:11 crc kubenswrapper[4715]: E1210 10:47:11.504595 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="481b5dc9-6323-4938-a300-45b98b999dd7" containerName="registry-server" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.504602 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="481b5dc9-6323-4938-a300-45b98b999dd7" containerName="registry-server" Dec 10 10:47:11 crc kubenswrapper[4715]: E1210 10:47:11.504620 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e73665f-3e61-4eda-8069-e05d39a614fc" containerName="copy" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.504627 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e73665f-3e61-4eda-8069-e05d39a614fc" containerName="copy" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.504831 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="481b5dc9-6323-4938-a300-45b98b999dd7" containerName="registry-server" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.504859 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e73665f-3e61-4eda-8069-e05d39a614fc" containerName="gather" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.509807 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d6b1960-9fec-49da-878c-afcf2a435ad2" containerName="collect-profiles" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.509902 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e73665f-3e61-4eda-8069-e05d39a614fc" containerName="copy" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.512065 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.542219 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7w2x7"] Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.597420 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-catalog-content\") pod \"redhat-operators-7w2x7\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.597479 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-utilities\") pod \"redhat-operators-7w2x7\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.597735 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qczf9\" (UniqueName: \"kubernetes.io/projected/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-kube-api-access-qczf9\") pod \"redhat-operators-7w2x7\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.699818 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qczf9\" (UniqueName: \"kubernetes.io/projected/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-kube-api-access-qczf9\") pod \"redhat-operators-7w2x7\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.700086 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-catalog-content\") pod \"redhat-operators-7w2x7\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.700141 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-utilities\") pod \"redhat-operators-7w2x7\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.700870 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-catalog-content\") pod \"redhat-operators-7w2x7\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.700890 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-utilities\") pod \"redhat-operators-7w2x7\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.718597 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qczf9\" (UniqueName: \"kubernetes.io/projected/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-kube-api-access-qczf9\") pod \"redhat-operators-7w2x7\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:11 crc kubenswrapper[4715]: I1210 10:47:11.841129 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:12 crc kubenswrapper[4715]: I1210 10:47:12.347227 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7w2x7"] Dec 10 10:47:13 crc kubenswrapper[4715]: I1210 10:47:13.015657 4715 generic.go:334] "Generic (PLEG): container finished" podID="4dda0699-3abf-4c3d-a4e9-2dfea51da5bc" containerID="55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e" exitCode=0 Dec 10 10:47:13 crc kubenswrapper[4715]: I1210 10:47:13.015752 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w2x7" event={"ID":"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc","Type":"ContainerDied","Data":"55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e"} Dec 10 10:47:13 crc kubenswrapper[4715]: I1210 10:47:13.015964 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w2x7" event={"ID":"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc","Type":"ContainerStarted","Data":"a80d2c1f610d5bafec4ad0b9bbab926973ae80951e0ba094033662e48e13aee0"} Dec 10 10:47:15 crc kubenswrapper[4715]: I1210 10:47:15.043600 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w2x7" event={"ID":"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc","Type":"ContainerStarted","Data":"4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f"} Dec 10 10:47:17 crc kubenswrapper[4715]: I1210 10:47:17.062372 4715 generic.go:334] "Generic (PLEG): container finished" podID="4dda0699-3abf-4c3d-a4e9-2dfea51da5bc" containerID="4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f" exitCode=0 Dec 10 10:47:17 crc kubenswrapper[4715]: I1210 10:47:17.062450 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w2x7" event={"ID":"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc","Type":"ContainerDied","Data":"4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f"} Dec 10 10:47:18 crc kubenswrapper[4715]: I1210 10:47:18.074262 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w2x7" event={"ID":"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc","Type":"ContainerStarted","Data":"d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3"} Dec 10 10:47:18 crc kubenswrapper[4715]: I1210 10:47:18.098391 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7w2x7" podStartSLOduration=2.61076432 podStartE2EDuration="7.098370723s" podCreationTimestamp="2025-12-10 10:47:11 +0000 UTC" firstStartedPulling="2025-12-10 10:47:13.01747521 +0000 UTC m=+4395.761021461" lastFinishedPulling="2025-12-10 10:47:17.505081613 +0000 UTC m=+4400.248627864" observedRunningTime="2025-12-10 10:47:18.09687786 +0000 UTC m=+4400.840424121" watchObservedRunningTime="2025-12-10 10:47:18.098370723 +0000 UTC m=+4400.841916974" Dec 10 10:47:21 crc kubenswrapper[4715]: I1210 10:47:21.841683 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:21 crc kubenswrapper[4715]: I1210 10:47:21.842326 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:22 crc kubenswrapper[4715]: I1210 10:47:22.891612 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7w2x7" podUID="4dda0699-3abf-4c3d-a4e9-2dfea51da5bc" containerName="registry-server" probeResult="failure" output=< Dec 10 10:47:22 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 10 10:47:22 crc kubenswrapper[4715]: > Dec 10 10:47:31 crc kubenswrapper[4715]: I1210 10:47:31.998288 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:32 crc kubenswrapper[4715]: I1210 10:47:32.044234 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:32 crc kubenswrapper[4715]: I1210 10:47:32.255093 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7w2x7"] Dec 10 10:47:33 crc kubenswrapper[4715]: I1210 10:47:33.207031 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7w2x7" podUID="4dda0699-3abf-4c3d-a4e9-2dfea51da5bc" containerName="registry-server" containerID="cri-o://d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3" gracePeriod=2 Dec 10 10:47:33 crc kubenswrapper[4715]: I1210 10:47:33.660039 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:33 crc kubenswrapper[4715]: I1210 10:47:33.823139 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-utilities\") pod \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " Dec 10 10:47:33 crc kubenswrapper[4715]: I1210 10:47:33.823316 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qczf9\" (UniqueName: \"kubernetes.io/projected/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-kube-api-access-qczf9\") pod \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " Dec 10 10:47:33 crc kubenswrapper[4715]: I1210 10:47:33.823432 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-catalog-content\") pod \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\" (UID: \"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc\") " Dec 10 10:47:33 crc kubenswrapper[4715]: I1210 10:47:33.824213 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-utilities" (OuterVolumeSpecName: "utilities") pod "4dda0699-3abf-4c3d-a4e9-2dfea51da5bc" (UID: "4dda0699-3abf-4c3d-a4e9-2dfea51da5bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:47:33 crc kubenswrapper[4715]: I1210 10:47:33.829130 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-kube-api-access-qczf9" (OuterVolumeSpecName: "kube-api-access-qczf9") pod "4dda0699-3abf-4c3d-a4e9-2dfea51da5bc" (UID: "4dda0699-3abf-4c3d-a4e9-2dfea51da5bc"). InnerVolumeSpecName "kube-api-access-qczf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 10:47:33 crc kubenswrapper[4715]: I1210 10:47:33.924846 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 10:47:33 crc kubenswrapper[4715]: I1210 10:47:33.924879 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qczf9\" (UniqueName: \"kubernetes.io/projected/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-kube-api-access-qczf9\") on node \"crc\" DevicePath \"\"" Dec 10 10:47:33 crc kubenswrapper[4715]: I1210 10:47:33.947824 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4dda0699-3abf-4c3d-a4e9-2dfea51da5bc" (UID: "4dda0699-3abf-4c3d-a4e9-2dfea51da5bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.028078 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.218864 4715 generic.go:334] "Generic (PLEG): container finished" podID="4dda0699-3abf-4c3d-a4e9-2dfea51da5bc" containerID="d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3" exitCode=0 Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.218907 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w2x7" event={"ID":"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc","Type":"ContainerDied","Data":"d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3"} Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.218958 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w2x7" event={"ID":"4dda0699-3abf-4c3d-a4e9-2dfea51da5bc","Type":"ContainerDied","Data":"a80d2c1f610d5bafec4ad0b9bbab926973ae80951e0ba094033662e48e13aee0"} Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.218985 4715 scope.go:117] "RemoveContainer" containerID="d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.219063 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7w2x7" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.240887 4715 scope.go:117] "RemoveContainer" containerID="4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.275081 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7w2x7"] Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.280245 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7w2x7"] Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.282675 4715 scope.go:117] "RemoveContainer" containerID="55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.330730 4715 scope.go:117] "RemoveContainer" containerID="d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3" Dec 10 10:47:34 crc kubenswrapper[4715]: E1210 10:47:34.334446 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3\": container with ID starting with d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3 not found: ID does not exist" containerID="d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.334489 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3"} err="failed to get container status \"d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3\": rpc error: code = NotFound desc = could not find container \"d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3\": container with ID starting with d67f010c0bb4ecea2c7ba204871755c8341b2a8d7e1ad5189ddd62b44eeaf4b3 not found: ID does not exist" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.334515 4715 scope.go:117] "RemoveContainer" containerID="4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f" Dec 10 10:47:34 crc kubenswrapper[4715]: E1210 10:47:34.334715 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f\": container with ID starting with 4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f not found: ID does not exist" containerID="4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.334740 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f"} err="failed to get container status \"4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f\": rpc error: code = NotFound desc = could not find container \"4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f\": container with ID starting with 4ab3cb140f7382f15d88c3a69821c9ac934498888f0320a699f32156ecb8495f not found: ID does not exist" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.334754 4715 scope.go:117] "RemoveContainer" containerID="55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e" Dec 10 10:47:34 crc kubenswrapper[4715]: E1210 10:47:34.335697 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e\": container with ID starting with 55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e not found: ID does not exist" containerID="55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e" Dec 10 10:47:34 crc kubenswrapper[4715]: I1210 10:47:34.335767 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e"} err="failed to get container status \"55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e\": rpc error: code = NotFound desc = could not find container \"55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e\": container with ID starting with 55caafe87429c542c8b724dbd81ceaef3418cac9107bc2cec86c86403be86a7e not found: ID does not exist" Dec 10 10:47:35 crc kubenswrapper[4715]: I1210 10:47:35.618906 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dda0699-3abf-4c3d-a4e9-2dfea51da5bc" path="/var/lib/kubelet/pods/4dda0699-3abf-4c3d-a4e9-2dfea51da5bc/volumes" Dec 10 10:48:17 crc kubenswrapper[4715]: I1210 10:48:17.714693 4715 patch_prober.go:28] interesting pod/machine-config-daemon-7pbhk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 10:48:17 crc kubenswrapper[4715]: I1210 10:48:17.716276 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7pbhk" podUID="8b26b009-53b8-4260-88a0-0f9f9553d676" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116250001024434 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116250001017351 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015116236670016514 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015116236671015465 5ustar corecore